> It's the fact that Facebook is hemorrhaging its best people as those on the inside finally start waking up to the fact that Zuckerberg's vision of Facebook as a good thing for the world was a lie.
I would say that if these people are the best, they it couldn't have taken them this long to figure out what's obvious to so many of no.
Having been in the academia I can tell you that from my experience there's two main reasons why scientists are often reluctant to share their code:
A) embarassement at having other people see how bad it is. Not necessarily wrong, but bad in other ways.
B) wanting to keep an advantage by continuing to publish on top of the work already done, whereas anyone else wanting to get in on the same idea would have to first re-build the original work.
Scientists have to have the truth of their malpractice rubbed hard against their face before they admit that they have a problem. Since Aristotle science has be wrought with what essentially amounts to bullshit and it only ever changes once so many people call bullshit so frequently that a new standard develops.
Not pre-publishing a research thesis / methodology is bullshit.
Not publishing data is bullshit.
Not publishing exact code (and runtime, etc) is bullshit.
Not publishing when research is complete is bullshit[0].
Not publishing is bullshit. I don't care who your funder is, it should be illegal for funders to selectively publish research. It's obviously one sided propaganda. The only time not publishing makes sense is when there are national security or global security ramifications to the research or if funding needs to be lined up for patents.
[0] I've seen a PhD candidate delayed 4 years from publishing because her supervisor was trying to coordinate a team to publish with research along the same vein all at once to make it a groundbreaking set of discoveries. It's basically p-hacking by some other name.
Not academia, but in finance, I remember an email interaction that was very nearly company-wide at a large hedge fund. In both our production & development environments, all generated core files on our Linux servers were configured to be dumped to a shared NFS mount (ostensibly to ease debugging). Each environment had around a 1 or 2 TB mount. Teams were expected to clean up their cores when they were no longer needed (useful for debugging those heisenbugs). Nearly all of the code was C++. Think 10s to 100s of millions of lines of C++ that had to work together spread across hundreds of services, countless libs that were used to interconnect.
Anyways, we were working on a fairly large company-wide effort. We were migrating from 32-bit to 64-bit, had 2 OS upgrades (new version of RHEL, migrating from Windows XP to Windows 7), 2 compiler upgrades (I forget the gcc versions - it came with the upgrade, and on Windows VS2003 to VS2008). Because of the coupling of the libs & services, both the Linux & Windows sides had to be released at the same time. We could update the Windows boxes at anytime and run the old software on it, but we basically couldn't develop the old software on Win7 as VS2003 wasn't compatible (it could be made to work, but the IDE or compiler or linker may fail randomly and hard). This is just to explain the scope of the effort we were making.
Back to why I mentioned the core files. To anyone that's done it, it's obvious the magnitude of effort doing this all at once is. There will be bugs. Specifically a metric-shit-ton of them. Those core files were needed. They stopped being generated because our shared core space filled up. The culprit? A PhD quant running some model in Python using some C/C++ extensions kept crashing Python, each one dropping a multi GB core file. This one quant was single-handedly using over half of the shared space. When confronted/inadvertently shamed in a development-wide email (we sent out usage breakdowns per user/project when a usage threshold was exceeded), his response was golden: "What are these core files, and how do I stop them from being generated?" Uniform response from everyone else: "Fix your damn code!" Mind, at the time, this was in front of ~600 developers.
The kicker: the reason this whole effort was being made? Because someone sent the CEO an XLSX spreadsheet in 2008 (the exact year may have been later, but immaterial to the story), and still being on WinXP & Office 2003, he couldn't open it. So...down the rabbit hole we went.
B is utterly unacceptable if you're funded by public money. As a matter of fact, I don't understand why public research grants don't come with a binding obligation to make all results available, including source code.
While still in academia I wrote some code to extract some data from images. It took about a week before I was happy with it. We then decided to make it available for other people. Polishing it until somebody else could use it with reasonable effort took several months and a team of students. That is time that could have been spent writing more papers. It is not at all clear whether enough people actually use the software to justify the effort.
Researchers aren't paid (and usually aren't trained) to write software that can be installed on more than one computer. I don't like that situation either, but that how it currently is. Research code also suffers from accelerated bit rot. The dependencies are often research code themselves and projects are abandoned when the results are published.
And what made you thing that "I was happy with it" is enough from the point of view of the pursuit of scientific truth (e.g. how did you deal with the problem of confirmation bias)? More than half a century of experience with software engineering has told us the hard way that it's near impossible to write bug-free code. And where this happens (e.g. aviation software) it takes extreme dedication and resources.
There is a reason that Alan Turing invented program verification in the 1940s [1].
Let me finish with a quote from
computing pioneer M. Wilkes [2]: "I well remember when this realization first came on me with full force. The EDSAC was on the top floor of the building and the tape-punching and editing equipment one floor below. [...] It was on one of my journeys between the EDSAC room and the punching equipment that "hesitating at the angles of stairs" the realization came over me with full force that a good part of the remainder of my life was going to be spent in finding errors in my own programs."
The hard part wasn't getting the bugs out (we didn't find anything substantial), it was making it run on a machine different from my laptop, making it take non-hardcoded data and parameters, cleaning the code up and documenting sufficiently that we could reasonably understand it after not looking at it for half a year, and packaging it in such a way that users wouldn't have to spend half a day installing it.
I don't think you can reasonably expect software that extracts features from photographs to be proven correct. What would the specification even look like? How many man years would you spend on verifying OpenCV and the two dozen other dependencies you rely on? It's not like all math papers provide machine checked proofs and that would probably be an easier endeavor.
You label a few photos yourself and then check if the labels still correspond in the newer versions of the program.
Then when you edit any parts of the code and the test suite breaks - you can see that some of the invariants had been broken.
This is even more important for python, as at least having a simple test suite can already tell whether your environment is sane. And that's about 90% of the newcomer's time saved.
I agree that formal verification in this problem space is currently not feasible (even if we ignore the oracle problem: where to get the spec from). But let's not make the perfect the enemy of the good. Even a basic reproducible test suite (e.g. running the software on some very basic pictures with a known result, and running against similar software) in an automated fashion, would help. Almost no scientific software I've seen get even those basics right.
I would rather the publication give a clear enough description that someone with time and resources could reproduce the software based on that description and achieve the same results. I think that would actually foster more innovation and collaboration as people implementing the software from the publication will necessarily become much more familiar with the work than they would have otherwise.
Not really answering the question, which was "how do you know it works?"
I would contend that if you have proven to yourself that your code works and are therefore capable of proving it to other folks (via e.g. solid testing), you should not be ashamed of the spit and glue.
It is research code, we all know what that looks like. But if, on the other hands you haven't proven to yourself it works, then it's definitely something to be ashamed of - scientifically speaking.
It’s much the same reason why you wouldn’t go to work wearing a days old, stained shirt and ripped pants — yeah, you’re more than likely going to work just as hard and as well than if you were wearing clean clothes, but that’s not going to stop people from judging you based on your appearance.
Most academics write code to just work. Not work well, or to be generalized, or to be efficient, just work. And while that’s absolutely fine, as your results being reproducible from the code is all that really matters, a lot of people don’t see it that way and will only see code slapped together haphazardly and dismiss you because of it.
This article shows very clearly why that does not work. The Princeton team could not reproduce the Berkeley results, yet the inaccessibility of the code meant that the latter persisted as a road-block for almost a decade.
Imagine if this reproducibility excuse were applied to experimental results and technique: we don't have be careful or explain in detail what we are doing, as reproducibility will take care of any errors. One consequence would be that, as the current state of knowledge became less certain, it would become less clear what to do next.
I understand the embarrassment, I am sometimes embarrassed as well looking at quickly-put-together code
I do appreciate papers that give you either a reference implementation or at least enough details on how it was done. I spent too many hours on papers that didn't trying to reproduce their claimed results...
Usually testing. Just because code works flawlessly doesn't mean it isn't embarrassing; and just because you are proud of the code doesn't mean it's bug free.
I don't disagree from your overarching point, but just a nitpick:
I mostly disagree when you say that Nazis were socialists. Of those people who supported the Nazis, the vast majority did so in part because the Nazis marketed themselves to those people as socialists. On the other hand, Hitler marketed the Nazis to "the industrialists" as enemies of the communists: in Germany too communism and socialism was bundled together. So I would say it's more accurate that Nazism was a bunch of marketing ideas that Hitler et al used to grab power.
Now, you could say the same thing about the soviets, that their ideology was just a tool for grabbing power. They might have flipfloped on other topics, but I don't think they flipfloped on the specific topic of socialism.
They maintained a socialist state, financing the healthcare of all with the contributions (not symmetrical) of society. Not surprising. Spain’s healthcare system (arguably one of the best public ones) was started by Franco, a fascist ditactor.
They also did some other pretty socialist-things with ecology
> is a matter of considerable debate to say the least.
Correct. Ironically the person you're responding had an issue with the black and white claim that "socialism=communism" and wanted to add nuance, and did that by presenting a black and white claim that "the nazis were socialists", which is a claim that could use a lot more nuance.
I'm not so sure I buy this article. It seems to spend quite a bit of time defending claims that the Nazis were socialist from the obvious modern-day attacks of Trump supporters, which is probably not the way to get the most historically-valid ideas.
I spent a (tiny) bit of time looking for other articles on this topic, and it seems like one of the main arguments goes something like "Nazis were not socialists - they were racists". However, I really fail to see what one thing has to do with the other. The other main argument is much better - that the Nazis were using the word "socialist" in their name as a cynical political ploy, and not really an expression of their "true beliefs".
For my part, I have to say I was quite surprised when recently reading "Rise and Fall of the 3rd Reich", which talked a bit about the Nazis actual policies towards the economy. I have to say, they certainly sounded very socialist to my (libertarian-leaning) ears. I mean, they took major control of the economy, dictating work hours and salaries, among other clearly socialist/communist things that I can't quite remember anymore.
I think it's a pretty interesting question, and as others have said, "Socialism" isn't so well-defined. But looking at actual Nazi policies instead of motivations and rhetoric seems to me to be table-stakes in actually understanding the issue.
I'm not employing a specific strategy. A better wording would be discovering covered calls. I already own the shares so the downside risk is inherent. Betting against upside is great in these volatile markets, and given enough capital for the stock, I can buy 100 more if I get close to break-even price on the upside. For the last two weeks I sold calls for TQQQ.
For our American friends, this is like finding that the Democrats screwed over Bernie Sanders during the election campaign. Sounds familiar?
There's a difference, which is that here Corbyn actually became the Labour party leader. However despite that he's still portrayed as an outsider by the establishment and the media.
The two situations really aren't that similar at all. You don't have to make everything into an opportunity to bring up the 2016 election all over again.
You’re right, the difference was that Labour was unable to prevent Corbyn from becoming the party leader, while the DNC was already decided on anointing Hillary regardless of her actual popularity or electability and was able to prevent all opposition.
- Corbyn is the leader of the Labour Party. Neither Hillary nor Bernie have that position in the Democratic party, because it doesn't exist. A party's presidential candidate is not the same as their party leader. And the method by which each is elected is radically different.
- Corbyn was already leader when these events happened. This is the equivalent of Bernie winning the primary (which as mentioned above is already a poor analogy) and then being lied to about what the DNC was doing in outreach for the presidential election. So, not something that mirrors reality at all.
- Bernie Sanders was not a member of the Democratic party before seeking to become it's presidential candidate. Corbyn has been a member of the Labour party for decades.
- Corbyn represents the resurgence of a relatively old wing of the Labour party, not the beginning of a new movement like Sanders.
Basically every part of this dynamic is different. They're both very interesting to study in their own right, and trying to conflate the two to make a political point does nothing justice.
In both cases, the party establishment tried to prevent a candidate on the left from gaining power through deceitful means. They tried to rig things against Bernie and Sanders.
That’s the comparison being drawn here. Does that make sense?
(Yes, the details are different as the mechanics of these party systems are very different.)
To the extent that the comparison is so absurdly simplistic that it doesn't do justice to either of the situations it's trying to describe, sure, it makes sense.
Staff within the Labour Party running ads to deceive their party leader into believing they were running a more progressive message than they actually were is literally nothing like anything that happened to Bernie Sanders in 2016.
I think this is more like finding out that the republicans screwed over Donald Trump. There's a strong case to be made for Corbyn being in the mold of Trump, but from the left instead of the right.
In the American case, private emails in the dnc were talking about preferences of officials. Don’t forget that Bernie Sanders isn’t even a Democrat! So the officials had a preference for a Democrat to win, whatever.
No one has pointed to actual harm caused. Ultimately Bernie lost in the primaries even before the super delegates kicked in. Probably didn’t help he didn’t campaign in the south which is a requirement for winning.
Bernie lost fair and square and no one has alleged specific wrong doing. Just vague “but emails” - this is worse than circumstantial evidence.
So yeah I’d say a far reaching ad campaign to deceive is not at all the same.
Seriously? That lawsuit certainly does not demonstrate at all, in any way, that any of the primaries were "rigged." It shows what is and was widely known: the DNC had a preference for one candidate.
It's extraordinary that two years after the fact so many people still do not understand what happened even though the facts are widely known and documented. This is the power of propaganda but still it's always a bit disturbing to see.
There's quite a few in London, generally on high streets and in stations rather than residential shopping areas - it's more of a boulangerie/patisserie/cafe than a bakery per se. e.g. Fleet Street - https://www.google.co.uk/maps/@51.5142772,-0.1074076,3a,75y,...
> It's not just the administration. Every time the EU has levied fines against Google a lot of Americans on HN claim that the EU is doing it because they're jealous of how successful American businesses are.
Currently using bitcoin for payments changes nothing. Yes, the transfer of "currency" are decentralized, but to cash it out (to pay for food, rent, etc) you still need to use exchanges which are centralized.
A) Regardless of whether you're right or wrong, that's just a discussion about labels. Call it whatever you want, my point still stands: the transfers are decentralized but the exchanges are centralized.
B) Even though my point does not depend on whether you're right or wrong, you're wrong. I have USD in my bank but I have to cash them out into GBP. So according to your "reasoning", USD is not a currency? Of course it is, you're just talking nonsense.
Either a currency is something you can buy things with, or it's a label with exchange rates. In the first case it can't be something that needs to be cashed out. In the second case you can't buy something and the thing you buy other things with is money.
You're usage of Bitcoin, money, GDP, USD and currency is willynilly.
I would say that if these people are the best, they it couldn't have taken them this long to figure out what's obvious to so many of no.
No, these are rats leaving the ship.