These thousands of words of repeating claims with weird non sequitur digressions seem to amount to "I wish Python used hard-to-enter unicode characters instead of words on normal keyboards" as far as I can tell.... because human brains, apparently, cannot make sense of the two character symbol `:=` but could somehow easily comprehend the exact same operation using the back arrow that I have no idea how to enter here.

And yes, I know that Martin said there was some complex chord of keys in one particular IDE that would enter that arrow character.  I don't remember what the combo is, but I'm sure it's possible to learn it.

FWIW, as I often point out is similar threads, I actually use vim's conceal plugin so that my Python code looks like it has funny characters inside it.  But it doesn't really, the screen just shows those for certain character patterns, but the beautiful ASCII gets saved to disk as regular Python.  Martin could easily use that plugin (or something similar for his own editor) to visually transform ':=' into <funny back arrow>.

Just install something like that, and know that the key strokes to enter funny-back-arrow are ':' followed by '='.  Simpler than that chording in whatever APL IDE even.

On Sun, Nov 10, 2019 at 3:53 PM Martin Euredjian via Python-ideas <python-ideas@python.org> wrote:
> This has nothing to do with representation or input via text

It does, it's an extension of the reality that, after so many decades, we are still typing words on a text editor.  In other words, my comment isn't so much about the mechanics and editors that are available as much as the fact that the way we communicate and define the computational solution of problems (be it to other humans or the machine that will execute the instructions) is through typing text into some kind of a text editor. 

When I say "text" I mean "a through z, numbers and a few symbols that were found on mechanical typewriters in the 1960's".  My shorthand for that is ASCII, which isn't quite accurate in that the set symbols contained in the sets where the most significant bits are "000" and "001" (7 bit ASCII) are not used other than CR, LF and HT. 

So, for the most part, programming, for the last 60 years or so --over half a century-- has been limited to the characters and symbols found on a 60 year old typewriter.

For some reason this evokes the lyrics from a Pink Floyd song, "Got thirteen channels of sh** on the T.V. to choose from".  The advances in computation since the 1960's have been immense, and yet we pretend that it is OK to limit ourselves to a 60 year old keyboard in describing and programming the next generation of AI systems that will reach unimaginable levels of complexity, power and capabilities.  

As I have mentioned in another comment, having had this experience, I fully understand how people who do not have the benefit of having communicated with computers, not just symbolically, but through a very different paradigm as well, simply cannot see what I am describing.  It's hard to find an analogy that can easily represent this without some common shared perspective.  I found that music can be that tool.  Of course, that requires classical training at a level sufficient enough to, for example, read and "see" the music when presented with a score.

Now, it's easy to say "I can do that" when presented with something like this and maybe have a rudimentary understanding of it:

https://www.youtube.com/watch?v=MeaQ595tzxQ 


It is something quite different when presented with something like this, without a "play" button, even if annotated:

http://buxtonschool.org.uk/wp-content/uploads/2017/04/Annotated-Bach-Brandenburg-score.pdf?LMCL=ruk_U7 

I have found that trying to explain the value of true notation to people who lack the experience and training is always a losing proposition.  I'm already regretting having started this thread, simply because I know how this works.  Frankly, it's almost like trying to engage with a religious person while trying to discuss the lack of evidence for the existence of supernatural beings.  They "know" what they "know" and it is a very rare case that someone is actually going to get out of that box and comprehend what you are saying.

BTW, there are some interesting efforts out there, like this:

https://www.youtube.com/watch?v=1iTPLgfmFdI

Once you dig into these truly interesting examples you end-up discovering that notation still has a significant edge.


-Martin


_______________________________________________
Python-ideas mailing list -- python-ideas@python.org
To unsubscribe send an email to python-ideas-leave@python.org
https://mail.python.org/mailman3/lists/python-ideas.python.org/
Message archived at https://mail.python.org/archives/list/python-ideas@python.org/message/T55CUZ6AVDK5LFZFHWJRG2RTNLEYSSTF/
Code of Conduct: http://python.org/psf/codeofconduct/


--
Keeping medicines from the bloodstreams of the sick; food
from the bellies of the hungry; books from the hands of the
uneducated; technology from the underdeveloped; and putting
advocates of freedom in prisons.  Intellectual property is
to the 21st century what the slave trade was to the 16th.