• 0 Posts
  • 7 Comments
Joined 1 year ago
cake
Cake day: June 30th, 2023

help-circle
  • Hey, looks neat.

    Hope you’re open to constructive criticism - I’d take a look at adding some production value to your how it works video.

    I work with a lot of martech folks that do product videos (I’m not selling you something) and I’d recommend a super straight forward marketing video that shows how easy the product is to use and share videos with.

    Get literally any budget microphone and record your audio voiceover VERY clearly in a closet and lay that over a simple workflow for capturing a video with snapify then sharing it. Add some royalty free background music at low volume and it’ll help sell this for you significantly more than your current video is doing.


  • I mostly agree with you but think it’s important to clarify that even with machine learning many humans can be replaced.

    To extend your metaphor, that library has always had a bunch of clerks sitting inside of it. They’ve been handling requests, finding books, and organizing them into a system that works to best serve that information.

    Now with machine learning, instead of having all of those clerks making the library run smoothly, they’ve effectively replaced 99% of all of the humans with an organizational system that serves content and helps find books even faster than a human would be able to.

    Slightly deeper: this machine learning replacement can also now mix and match bits of content. The human system before might have a request that looks like this - “I want information on Abrahamic Religion in Western Culture” so they’d gather up a ton of books and pass them to the person that requested info.

    In the new replacement system, the request could take bits and pieces from all of those books and present a mostly comprehensive overview of Abrahamic Religion in the West without having to run and fetch all of the books.

    Deeper yet, and the scary iceberg - today, someone still needs to write all of those books and we as a society tend to trust information gotten from those books (cited sources and all that) so humans are safe as the content authors right? We’ve basically just made a super efficient organizational and content delivery system. But as we start to trust the new system and use it more, we’re potentially seeing the system reference its own outputs as opposed to the source material…which creates a recursive, negative feedback loop.

    We still need human content creation today, but the scary part (IMO) is when we treat these LLMs as generative general AI. The LLMs are fallible and can be incorrect and often hallucinate - so when most people start blindly trusting these systems (they already do - look no further than general confusion on the terms AI and machine learning and LLMs), we’re going to get increasingly further away from new knowledge generation.