Vision concept: the internet without physical input devices

Cataloguing this under: “interesting ideas which don’t quite fit the broad vision concept but which I thought worth writing down” :slight_smile:

Many of the tasks we (or at least I) regularly use computers and the internet for fall under some form of possible Anoma application:

  • booking flights (at present: Google Flights, Skyscanner; Anoma: Modular Logistics)
  • booking a car (at present: Uber, Bolt; Anoma: Modular Logistics)
  • sending a bank transfer (at present: N26; Anoma: Kudos)
  • messaging (at present: email, Signal, Slack; Anoma: Multichat)
  • information retrieval (at present: Wikipedia, Google; Anoma: Knowledge Graph)

Typically, the interfaces for these kinds of applications, at present, consist of:

  • graphical (touch) and symbolic (keyboard, mouse) user input
  • bespoke visual display

Personally, I often find these interfaces obtrusive and alienating:

  • I must fetch a physical device to perform the input and view the display. This often pulls me out of the flow of some other activity (e.g. a call with friends planning a trip).
  • User input requires translating a desired semantics (e.g. “I want the cheapest flight available from X to Y on dates … not on a recent Boeing aircraft”) to specific input syntax (various buttons and form fields in Skyscanner, for example).
  • Only semantics that the programmer of the user input system has specifically chosen to support are supported, even if data is available for more expressive semantics on the “backend”.
  • I cannot easily use multiple applications (e.g. booking flights, hotel, festival tickets) in unison - I have to visit different interfaces for each, and the actions themselves cannot be performed atomically.
  • Information is often split across protocols for arbitrary reasons - e.g. I would naturally index my communications by counterparties and topics of discussion, but separate communications protocols (Slack/Signal/e-mail/etc.) introduce further (undesired) segmentation.
  • Available display options are typically limited, application-specific (cannot combine data across applications), and constrained by the imagination of the programmer.

With the combination of approximately:

  • Anoma as a distributed operating system,
  • A standard library of data types (e.g. Datetime, Location),
  • Key Anoma applications (Kudos, Multichat, Modular Logistics, Knowledge Graph),
  • Standard ontologies for objects (cars, airplanes, messages, …), and
  • Local LLMs

… one can imagine a very different set of interfaces to these kinds of applications. In particular, graphical and symbolic input can be replaced entirely with natural language (spoken or written), and bespoke visual display can be replaced partially with natural language output and partially with synthesized views. Using natural language I/O and synthesized views removes the need for custom, per-application input and outputs views and allows applications to be seamlessly combined from the user’s perspective (based on semantics - e.g. “I want a flight and a hotel”).

cc @apriori curious for your take here

2 Likes

Anoma fits really well with this future. In the last year or so of travel often when I talk to people about intents for long enough, LLMs come up—as a way to process user intents. This is interesting because there is clearly excitement about processing intents in natural language (text and voice dictation).

  • One cool part of this future is that it can remove the need for users to look at screens for application use beyond interacting with text. For example, If I could just have a conversation with an LLM that would help me book my plane, hotel, and train ticket without having to use a mobile application it would be a significantly better UX. It is somewhat of a return to the past, booking travel arrangements with travel agents.
  • This is overall a nice way to contrast Anoma vs. existing systems (Web 2/Web 3)
  • Primitive example of natural language intents used to execute trades
1 Like

I like this vision a lot! It reminds me of a decentralized Amazon Echo. An interesting question is how to secure the system, as the LLM will have control over my cryptographic identity. Surely, there exists some research around this (e.g., here). Additionally, intents constraining what the LLM can do are a simple solution to this.

1 Like

Had an interesting conversation with the founder of HeyElsa yesterday. They are willing to cooperate and integrate intents into their system. I played around with their prototype and it’s actually quite interesting to have this type of interaction with an LLM. Happy to provide any further info about the conversation. Long story short, there are definitely opportunities for us to go down the Intents <> LLM route.

1 Like