Show HN: Multimodal perception system for real-time conversation

(raven.tavuslabs.org)

32 points | by mert_gerdan 4 hours ago

5 comments

  • ycombiredd 42 minutes ago
    Hmm.. My first thought is that great, now not only will e.g., HR/screening/hiring hand-off the reading/discerning tasks to an ML model, they'll now outsource the things that require any sort of emotional understanding (compassion, stress, anxiety, social awkwardness, etc) to a model too.

    One part of me has a tendency to think "good, take some subjectivity away from a human with poor social skills", but another part of me is repulsed by the concept because we see how otherwise capable humans will defer to "expertise" of an LLM due to a notion of perceived "expertise" in the machine, or laziness (see recent kerfuffles in the legal field over hallucinated citations, etc.)

    Objective classification in CV is one thing, but subjective identification (psychology, pseudoscientific forensic sociology, etc) via a multi-modal model triggers a sort of danger warning in me as initial reaction.

    Neat work, though, from a technical standpoint.

  • ashishheda 1 hour ago
    Wonder how it works?
  • Johnny_Bonk 44 minutes ago
    Holy
  • jesserowe 4 hours ago
    the demo is wild... kudos