Hacker News new | past | comments | ask | show | jobs | submit | yconst's comments login

>Our brains have this giant section dedicated to nothing but recognizing characters in our native written language. Sure we have a lot of brain power going to shape recognition in general, but character recognition is so incredibly trained in our heads that it should be our go to.

Yea, but you still have to read them, i.e. focus this "machinery" on a single word/sentence. So you're working in sequence when trying to identify a bunch of unrelated labels. On the contrary, you can probably differentiate an entire group of icons just by a glance in their direction (provided they are appropriately designed, of course).

Plus reading speed is pretty much the same no matter how many times you repeat the task. Recognition on the other hand improves with repetition.

If you're using only text, you're pretty much stuck with being able to differentiate your icons either by location or by actually reading them. Icons add another component to this.

Totally agree with the color-related comment though.

IANANS btw, so your mileage may vary.


I found the first comment to this article quite interesting:

"Marcus has a point - even is some of what is said about deep neural networks is incorrect (for instance, they can learn and generalize from very few example, one shot learning).

However, he got it wrong with the answer. The key for machines to reach the symbolic abstraction level is the way we train them. All training algorithms, supervised, unsupervised or reinforcement learning with LSTM rely on the assumption that there is an "utility function" imposed by some external entity. Problem is, by doing so, we are taking away their capacity [of the machines] to make questions and create meaning.

The most important algorithm for learning is "meaning maximization" not utility maximization. The hard part is that we cannot define what is meaning - maybe we can't, I'm not sure. That is something I will be glad to discuss."


You could always compare how much your perceived reality matches predictions from your mental model. If you look at the concept of 'schemas' in child development, you'll see children focusing on one thing over and over again, like 'containment'. Putting a block in a box, taking it out, putting it in, taking it out. Each time they're learning that when you put something inside something else, it's still there even if you can't see it, that you can pick them up together etc etc. They're slowly exercising the different ways they can interact with their environment, learning to model likely outcomes. Learn enough of these schemas and you can start making more complex plans to see what the world does in response.


> The hard part is that we cannot define what is meaning - maybe we can't, I'm not sure

I think you've correctly identified the million dollar question in AI, and that's why I believe the major breakthroughs in artificial general intelligence won't come from scientific researchers. They might even come from PG's dreaded literary theory types, ones who are also talented programmers (...Lit theory being maybe the only field that thinks about structuralism, deconstruction, how meaning is manufactured, etc. - things outright dismissed by the scientific community as useless, but which still hold potential to produce a 'meaning algorithm').


This is perhaps a bit philosophical, but assuming we had a "meaning maximization" function, what stops us from writing it as a loss function and using our current supervised machine learning frameworks?


Just because we can formulate an optimization objective, it is not guaranteed that we will find an algorithm that solves it within a reasonable amount of time. In case of humans, these objectives or preferred states are possibly very simple ones, like hunger and pain avoidance, reproduction and curiosity; and it is actually easy to write down an algorithm that optimizes these objectives (if you ignore how reality actually works): Simply try out all possible ways of reacting to the environment and choose the best one.

This works in theory, but in practice you only have a limited amount of chances to try something out (because of the arrow of time). This makes learning a necessity. You need to keep a record of all trials you have performed so that you can reuse this information later when the same situation reoccurs. How to do this in an optimal way is described by Bayes' theorem.

The key to AI will be a certain set of priors, biases and fixed function units that make this computationally tractable; we'll likely need things like invariance of the learned information to various changes so that it can be reused in different settings, segmentation of data coming from the world into episodes (hippocampus), attention, control (basal ganglia), mental rotation (cortex) and path integration (hippocampus, grid cells).


That's true, there are certainly many optimization objectives computationally intractable, or perhaps too abstract to be useful for learning.

However, I would argue the prior of Bayesian modeling can be just as nebulous and computationally intractable as an optimization objective. Like supervised learning, Bayesian modeling is just a tool.

I'm skeptical that we will reach AI through a deep understanding or modeling of the brain. Technology and computer science advances more quickly than the biological sciences, at least in recent times. You might argue a success in robotics like [0] is a motor control system. But they built this extending mathematical frameworks not being biologically inspired, and the big wins there didn't come from fixating on a learning framework or biological mimicry; just like humans learning to fly didn't come about from flapping wings like a bird. At some point we hacked an engine (invented for other purposes) onto a wing and came up with powered flight.

As an aside, only seeing input a limited number of times would likely improve your ability to find models that generalize as your model must be able to take these one off learnings and unify them in some way to achieve high training performance. With respect to human learning, a specific individual only has one chance, but nature has had many. We are only a selection of those chances that seemed to work well enough. There are many commonalities to existence that allow for this to work well in practice.

[0] http://groups.csail.mit.edu/rrg/papers/icra12_aggressive_fli...


Your agent may need a way to ask for a "kind" of training instance from the world in order to maximize meaning. Like maybe I've seen mammals, and now I need to see another kind of animal to maximize my understanding of the meaning animal. A human being—perhaps instinctively, or perhaps by some other force—has the curiosity to go find / pay attention to fish and birds. A kid can tell you he wants to go to the zoo.

A supervised machine learning framework can't tell the researcher what training instances it needs to see in order to improve its meaning. A supervised learning framework can't imagine where it might find that training instance, or describe what it may look like. A supervised learning framework never asks to go to the zoo.


Yes, in theory an offline supervised learner should never beat an online reinforcement learner. Adding a set of actions A that can be used to bias future examples in a predictable manner is certainly an advantage that will yield better convergence properties in almost all scenarios, simply because it lets you gain more information per observation.


I'm just curious: Can there be intelligence without some externally imposed utility function? For us humans it usually consists of access to (in decreasing order of importance) oxygen, water, nutrition, shelter, closure, etc. If an AI was free of all these constraints - meaning it had no predetermined utility function to maximize - then what is it supposed to do? It would have no motivation, no reason to learn or do anything. And if it had no goals which it tries to reach, how could we determine its performance or rather its intelligence?


Interesting point. However, I'd like to argue that there is a fine line between the conventional definition of a utility function and the ability of an organism to survive in an environment. The latter is truly open-ended, and not really externally imposed, i.e. different animals/people fare throughout life "optimising" completely different functions (closure?). A quote from a Kozma paper comes to mind: "Intelligence is characterized by the flexible and creative pursuit of endogenously defined goals". I believe this quite well summarises the open-ended nature of the task at hand. As I understand it, indeed, goals, rewards, risks, hazards, they are all in the game and shape the decision-making of the agent, it's "policy". But the way they are formalized for each individual and situation, well, is probably subject to constant redefinition itself.


> All training algorithms, supervised, unsupervised or reinforcement learning with LSTM rely on the assumption that there is an "utility function" imposed by some external entity. Problem is, by doing so, we are taking away their capacity [of the machines] to make questions and create meaning.

It's just a simplification. We are consuming the algorithm's output, rather than it being for internal consumption in a larger mind. So we supply the goal, where in a larger more natural mind the goal comes from other parts of the mind - still computing ultimately against a fixed utility function. That function being our human instinctual basis and the "firmware" of culture it's primed to load.


Well,

It's not just training that implies utility maximization, basic internal processes of a neural network, such as back propagation, are based on the utility function maximization. Back propagation is an approach that allows the network to do gradient descent on a utility function by propagating errors through the system.[1] Back propagation has been the way that deep networks have become tunable and effective. Without a utility function, it's hard to see how one would tune them.

And it's hard to think of what "mean maximization" could be but maximizing a utility function based on meaning.

And while you can construct a neural network that learns from a single example, the nn framework is fundamentally based on learning from multiple examples so such a construct is mostly meaningless.

But the point about machine learning being limited by the training methodology is good, it's just I'm pretty sure you'd different algorithms if you used a different methodology, the existing algorithms are used specifically because they fit the methodology.

[1]https://en.wikipedia.org/wiki/Backpropagation


Precisely the same question I have as well. Especially given that the energy you could store in a spring or coil could be much more, and you are not limited by one's lifting capacity or maximum height. Anyone have an idea?


"Do I dare to say this somehow shows the relative unimportance of design?"

When design is about cramming some animation in just about everything you can show on the screen, and about how fancy your colors are going to be – yes, I agree, it's relatively unimportant.

But real design is not about that at all. In fact, functionality has been part of the core debate in design, for at least a century now. Ask any designer – not only the ones dealing with UI, and they'll tell you.


"We could also take other data into account, like the user who submitted the article, and generate features indicating things like the karma of the user..."

I'm wondering how helpful could that be in prediction though? Would it actually help if I wish to predict how many upvotes my headline would get, and I add my karma as a feature? I think in fact such features would degrade generalisation performance, as they stand in like placeholders (when training), i.e. high karma users are correlated to higher probability for a "hit story".


It depends on how much data you can include in the prediction phase. If you train with karma earned right at the time of the article submission, and then predict with the exact same methodology, it will work fine. If you put in a "fake" karma value when predicting, then it will of course not work well.


So that's all great and trying to help and all, but.. a crowd that yells at the top of their lungs to a driver till he goes in tears (in place of actually guiding him to properly move the vehicle), and thereafter tries to lift the bus using nothing but brute force.. isn't this like creating a problem and then trying to solve it?


Wow, is that your take from the article? It just goes to show you can't please everyone.

I wonder if you would feel different if you were the one UNDER that bus?


The whole article tried to elaborate on the problem of crowds, did you read it thoroughly enough?


It does not elaborate on the "problem of crowds", at all. On the contrary, it offers a rather positive view to the whole spontaneous reaction thing. Which is great, don't get me wrong. However, in this case, I believe the solution could have been much much simpler to begin with. But that's just me :)


A related thought might be, once this is in place, how much work would it be so support Appkit (OS X) apps as well?


The weird thing is that such a product -- AppKit on Windows -- existed and was made by Apple itself. It was called "Yellow Box for Windows" and shipped as part of WebObjects at one point.

Now the tables are turned, and Microsoft is building an "iOS Yellow Box" of their own.


Indeed. Tell any ex-NeXT person circa 1996 that MS would be effectively building their own OPENSTEP for Windows and no one would have believed you.

Will be interesting see how they implemented it and whether they used Apportable's Foundation (which looks really well written).


I wonder if part of the reason for last year's Swift announcement was that Apple got whiff of Microsoft's Obj-C Cocoa clone.

Swift was obviously half-baked when it came out. It's still far from production quality IMHO. But as a deterrent for this kind of compatibility action, Swift's existence is quite effective.


- Swift uses the same compilers and frameworks

- Swift support s in Xcode and if Microsoft doesn't watch out their implementation might be miles ahead at their first try

- Swift 1.0 was a marketing 1.0, technically maybe a 0.2 release. I hope I don't need to touch it again in the next year even though I like the language principally


The fundamental issue is that glass is a blatantly obvious design, with no sophistication whatsoever, one that couldn't even get close to being charmingly anything. It's an ugly stick in front of some (admittedly ugly/geeky) glasses, with touchpad (!) and voice controls crammed together. No amount of functionality can account for the design and usability horror that this device has unleashed since it's conception.


Agreed. It's awfully awkward to operate.


But it will probably not scale linearly.


Seems $2k is just one of their milestones...

"development is split into milestones on which the features are discussed and decided upon in the mailing list with the help of the community.

a funding campaign is then setup for each next milestone which supports development and keeps the project up and alive."


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: