For those veteran linux people, what was it like back in 90s? I did see and hear of Unix systems being available for use but I did not see much apart from old versions of Debian in use.

Were they prominent in education like universities? Was it mainly a hobbyist thing at the time compared to the business needs of 98, 95 and classic mac?

I ask this because I found out that some PC games I owned were apparently also on Linux even in CD format from a firm named Loki.

  • ik5pvx@lemmy.world
    link
    fedilink
    arrow-up
    56
    arrow-down
    1
    ·
    6 months ago

    It was a struggle. You went to buy some device and you had to check it was not one of those windows-only ones. Modems were particularly bad, for example.

    You had to read the how-tos and figure things out. Mailing lists and newsgroups were the only places to find some help.

    You had to find the shop willing to honour warranty on the parts and not on the whole system, as they had no knowledge of Linux at all. But once you found them, you were a recurring customer so they were actually happy. You might even have ended up showing them memtest86!

    You would still be able to configure the kernel and be able to actually know some of those names, compilation would take several hours but it was a learning experience.

    You could interact with very helpful kernel developers and get fixes to test.

    You could have been the laughing stock of your circles of friends, but within you, you knew who’d have had the last laugh.

    And yes, Loki games had some titles working on Linux natively, Railroad Tycoon was one. Too bad they were ahead of the times and didn’t last much.

    • schizo@forum.uncomfortable.business
      link
      fedilink
      English
      arrow-up
      27
      ·
      6 months ago

      windows-only ones. Modems

      And, of course, they’d almost never actually SAY that on the box, so you had to see if you could look at what exact chip was on them and explain to a retail employee why you needed to look in the box, and that no, you certainly weren’t doing something sketchy, you just use Linux instead of wait why are you calling security…

    • eldavi@lemmy.ml
      link
      fedilink
      arrow-up
      10
      ·
      6 months ago

      compilation would take several hours but it was a learning experience.

      the first time i put gentoo on a g3 imac back in 2004; it took 3 days to compile everything and the computer got so hot that it warmed up the entire room like a space heater. lol

  • atzanteol@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    47
    ·
    edit-2
    6 months ago

    Everything was harder back then - even when using Windows. But you had to be a real masochist to run Linux.

    Computers were still quite new that most people had no real use for them beyond “work things”. Only nerds really used them for anything else. “Do you have an email address” isn’t a question you ask today.

    “Kids these days” don’t realize how easy they have it when it comes to just general comparability. There weren’t a lot of standards yet and vendors had proprietary drivers and offered no support AT ALL for “lye nux”. You had to do a ton of research and fiddling to figure out if there was any support for your specific version of a specific chip used by any peripheral you used. And then to discover that you had to patch your kernel to add a driver that somebody had bodged together. So now you were running your own fun custom-kernel so you could get full-duplex rather than simplex audio! But it works!

    Like - lets say today you want to buy an external IDE drive controller to plug in some old drives to for backup. You to to Amazon, search “USB external IDE enclosure” and buy the cheapest one you find. It probably works unless it’s defective. In '95 USB and Firewire were in their infancy so you would probably buy a serial or parallel port device. You would need to find whether Linux supported the specific version of the thing you wanted to buy, what tools there would be for it, etc. There was no standard “bulk storage device” driver that you could rely on or hope the vendor would implement. Even if you were an early adopter and got a USB or Firewire device it might have some “basic” functionality that works with OSS drivers but you couldn’t use all of it.

    Vendors back then also shipped their own software with things, not just drivers. It was always just the absolute worst crap that was buggy as shit. But it would do a lot of the heavy lifting in working with their device. Like any Creative Labs audio player you wanted to get working. Sure it used USB but it didn’t just mount as storage device, you needed to use the worst GUI ever put before mankind to use it (under Windows). Under Linux you had to find a specific tool that would support pushing/pulling media from it. These days it would just mount as a drive automatically and you’d use standard desktop tools to interact with it.

    Even with DOS/Windows you’d buy a game and as you came home from the store with it in a box wonder “will this work on my computer and how long will I need to mess with it?” I had to configure a specific CD-ROM driver to be used by DOS just to run Tie Fighter vs. X-Wing for example. Had a special boot floppy just for that game since that driver didn’t work with literally anything else I had.

    Hardware just generally didn’t “auto configure”. “Plug 'n Play” was still very much in its infancy and you often had to manually configure hardware and install special drivers just for a particular card or peripheral.

    IRQ 7 DMA 220. I probably just triggered some folks. If you were setting up a “Sound Blaster or compatible” then you had to know what interrupt it used (7) and what address it was on on the direct-memory bus (220). And you hoped there wasn’t a conflict with something else. If there was then there would be a DIP switch you could use to change the base memory address or IRQ from the default. But you were telling your software where to find the card.

    USB was a f’ing game changer for peripherals. Serial and parallel ports were so slow and obnoxious to use. Before that there was no real way to “probe” the bus to discover what was there unless you knew exactly what you were looking for (there’s no lsusb for serial ports). So you just guessed at the driver you need and “modprob foo” hoping it worked.

    It’s amazing what 20ish years of just developing standards has done.

    If you want a taste of that world I highly recommend LGR on YouTube. He’s mostly Windows focused but look for videos where he tries out “oddware” to see how often he has trouble getting things to work on period hardware using the vendor-supplied software even. Then multiply that by 100x for Linux. :-)

    • Cyborganism@lemmy.ca
      link
      fedilink
      arrow-up
      13
      ·
      6 months ago

      LoL!!! IRQ 5 DMA 220 for me. Had to manually adjust the jumper on the sound card.

      Fucking hell…

      • Jesus_666@lemmy.world
        link
        fedilink
        arrow-up
        6
        ·
        6 months ago

        Port 220.

        IRQ 5, port 220h, DMA 1 was what I used for my SoundBlaster 2.

        Later I used IRQ 5, port 220h, DMA 1, high DMA 5 for my SoundBlaster 16.

        • Cyborganism@lemmy.ca
          link
          fedilink
          arrow-up
          4
          ·
          6 months ago

          Do you think it’s worth getting a Sound Blaster card today? I’ve read you can get better sound effects in game. Can’t the on board audio chips do that now?

          • Jesus_666@lemmy.world
            link
            fedilink
            arrow-up
            7
            ·
            6 months ago

            I gotta be honest, I haven’t used a dedicated sound card since the Vista/7 era when EAX stopped being a thing and onboard sound could handle 5.1 output just fine. The last one I had was a SoundBlaster Audigy.

            These days the main uses for dedicated sound interfaces are for when you need something like XLR in/out and then you’ll probably go with something USB.

    • DAMunzy@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      10
      ·
      6 months ago

      I was reading your wall of text chomping at the bit to complain about IRQs and dip switches but you covered even that!

      Oh wait, you didn’t include having a math coprocessor daughter boards! I barely remember them but remember my dad building computers with them.

      I kinda wish I was a teen when the first computer kits were coming out. And phone phreaking.

    • Telorand@reddthat.com
      link
      fedilink
      arrow-up
      7
      ·
      6 months ago

      I wasn’t that into computers at that point in my life, but it was definitely a time where “computers” was a hobby, in the same way that restoring old motorcycles was/is a hobby. Sure, you might take it out for a spin every now and again, but a lot more time is spent tinkering than simply using.

      I’m constantly amazed by how much better the end-user experience is today, even just from 10 years ago. The installers are better, the pre-configured software and settings are more thoughtfully chosen, and now we’re at the beginnings of meaningful Linux gaming for non-hobbyists.

      We truly stand upon the shoulders of giants, and I look forward to the future.

      • atzanteol@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        6
        ·
        6 months ago

        Gaming has been the biggest change in the last 10 years or so. Mostly thanks to Steam. It’s easier to game on Linux these days than it is MacOS! It’s crazy.

    • aksdb@lemmy.world
      link
      fedilink
      arrow-up
      7
      ·
      6 months ago

      I remember buying a bunch of old HP ISA 100Mbit NICs to turn an old computer into a router/server combo. Naive as I was I put them all in and nothing worked. Turns out they were all configured to use the same IRQ (since they likely came from independent machines), and that caused them to overwrite each others settings… including the MAC adress. Thankfully I found some nice hacker that worked with these cards before and published a little C tool to rewrite their EEPROMs. I contacted him if he sees a chance to resurrect the cards and that saint indeed hacked the necessary features into his tool so I could rewrite the MAC adresses, change the IRQ one by one and ended up with a working network. Good times.

    • MonkderDritte@feddit.de
      link
      fedilink
      arrow-up
      6
      ·
      edit-2
      6 months ago

      Even if you were an early adopter and got a USB or Firewire device it might have some “basic” functionality that works with OSS drivers but you couldn’t use all of it.

      Oh, like scanners still.

      • atzanteol@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        9
        ·
        6 months ago

        Scanners and printers are one area of computing that have always sucked the most relative to other things. They’re better these days but they’re still the one thing I expect to fail on a regular basis.

    • arran 🇦🇺@aussie.zone
      link
      fedilink
      arrow-up
      3
      ·
      6 months ago

      This. However from about the release of knoppix and ubuntu things started looking and feeling a lot more like they are today. – I credit that to Knoppix for X & filesystem work and Ubuntu for setup and everything desktop.

      So even though late 90s it was tough, it was nothing like mid 90s. But by around 2004-2005ish the install and setup was substantially easier however the reputational damage still exists to today.

      I remember spending a lot of time in XFree86 config files, re-configuring it trying to figure out what works best on my monitor, and then the migration to XOrg. All good times.

      There was however a substantial amount of hype around Linux. It wasn’t quite what it is with AI. But you couldn’t read a magazine without encountering it in some way, but it was the type of hype were everyone knew of it but few people had anything to do with it.

      Another thing that hadn’t been mentioned is that there was a new distribution cropping up every day or so. (It felt like at least.) But this seems to back up that statement: https://en.m.wikipedia.org/wiki/File:Linux_Distribution_Timeline.svg

  • Blaster M@lemmy.world
    link
    fedilink
    English
    arrow-up
    37
    ·
    6 months ago

    Ah, yes, Linux around the turn of the century. Let’s see…

    GPU acceleration? In your dreams. Only some cards had drivers, and there were more than 2 GPU manufacturers back then, too… We had ATi, nVidia, 3dfx, Cirrus, Matrox, Via, Intel… and almost everyone held their driver source cards close to their chest.

    Modems? Not if they were “winmodems”, which had no hardware controller, the CPU and the Windows driver (which was always super proprietary) did all the hard work.

    Sound? AC’97 software audio was out of the question. See above. You had to find a sound blaster card if you wanted to get audio to work right.

    So, you know how modern linux has software packages? Well, back then, we had Slackware, and it compiled everything gentoo style back then. In addition, everyone had a hardon for " compiling from source is better"… so your single core Pentium II had to take its time compiling on a UDMA66-connected hard drive, constrained with 32 or 64 MB RAM. Updating was an overnight procedure.

    RedHat and Debian were godsends for people who didn’t want to waste their time compiling… which unfortinately was more common even so, because a lot of software was source only.

    Oh, and then MP3 support was ripped out of RedHat in Version 9 iirc, the last version before they split it into RHEL and Fedora. RIP music.

    As for Linux on a Mac, there was Yellowdog, which supported the PPC iMacs and such. It was decently good, but I had to write my own x11 monitor settings file (which I still have on a server somewhere, shockingly, I should throw it on github or somewhere) to get the screen to line up and work right.

    Basically, be glad Linux has gone from the “spend a considerable amount of time and have programming / underhood linux knowledge to get it working” to “insert stick, install os, start using it” we have now.

    • MonkderDritte@feddit.de
      link
      fedilink
      arrow-up
      3
      ·
      6 months ago

      In addition, everyone had a hardon for " compiling from source is better"

      I mean, optimization had more of an impact on the weak CPU’s back then, no?

      • Blaster M@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        1
        ·
        6 months ago

        That only matters if there’s anything to optimize by source compilation. If the program doesn’t have optimization features in the source, it’s wated time and energy.

  • The Zen Cow Says Mu@infosec.pub
    link
    fedilink
    arrow-up
    24
    ·
    6 months ago

    Way back in the early 90s I needed to use LaTeX for university. The dos version was awful and couldn’t handle large documents. So the options were (1) a nextcube for $$$$, (2) Nextstep 3.3 for PCs for $$$ (some faculty had this), or (3) linux. So I downloaded slackware on dozens of disks.

    You had to configure the kernel, which wasn’t too hard since the autoconfig walked you through it. The hardest part was setting up X11, which required a lot of manual config, and if you screwed up the timings you could destroy a CRT monitor. OpenStep was an option, so there was a moderately friendly windowmanager available.

    Learning Emacs was also fairly unpleasant, but that was the best option for editing TeX at the time.

    Everything would work, until it suddenly would break. But nonetheless I was somehow able to get that thesis done.

    Ugh, modern linux is SOOOOOOOOOOOOOOOOOO much better

    • Aceticon@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      6 months ago

      Just to add to this, early on there was no such thing as kernel modules, so you had to compile your own kernel with the hardware support you needed for anything beyond basic (if I remember it correctly, it was only basic processor stuff, keyboard and text mode VGA) hardware support.

    • JaxNakamura@programming.dev
      link
      fedilink
      arrow-up
      2
      ·
      6 months ago

      So I downloaded slackware on dozens of disks.

      This is no joke. When I downloaded Slackware in '95 or '96, it was over 100 3.5" floppies of 1.44 MB each. And there were still more available, those were just the ones I thought I’d need. And before you could even begin installing, each of those had to be downloaded, written and verified because floppies were not terribly reliable.

  • HarriPotero@lemmy.world
    link
    fedilink
    arrow-up
    24
    ·
    edit-2
    6 months ago

    Slackware and Red Hat were the two distros in use in the mid 90s.

    My local city used proper UNIX, and my university had IRIXworkstations SPARCstations and SunOS servers. We used Linux at my ISP to handle modem pools and web/mail/news servers. In the early 2000s we had Linux labs, and Linux clusters to work on.

    Linux on the desktop was a bit painful. There were no modules. Kernels had to fit into main memory. So you’d roll your own kernel with just the drivers you needed. XFree86 was tricky to configure with timings for your CRT monitors. If done wrong, you could break your monitor.

    I used FVWM2 and Enlightenment for many years. I miss Enlightenment.

    • mrvictory1@lemmy.world
      link
      fedilink
      arrow-up
      6
      ·
      6 months ago

      I used Enlightenment on Arch Linux for a year, in 2020-21. The PC had 4G ram and an HDD, Enlightenment was blazing fast. I could type enlightenment_start to a tty and reach a Wayland desktop under a second with 250M ram used total. E is still alive and kicking.

    • constantokra@lemmy.one
      link
      fedilink
      arrow-up
      2
      ·
      6 months ago

      How wrong did you have to be to break your monitor? Because I’m positive I got it very wrong a whole lot of times and never managed that.

      • cmnybo@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        5
        ·
        6 months ago

        By the late 90’s most monitors were smart enough to detect when sync speed was too far off and not try to display an image.
        It was the old monitors that only supported a single or fixed set of scan rates that you had to worry about damaging. Some could be very picky and others were more tolerant.

      • AnUnusualRelic@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        6 months ago

        I managed to make mine do some very worrying noises, but none of my monitors broke either, even though the bandwidth I based my calculations on was often kinda made up.

    • andrewth09@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      6 months ago

      If done wrong, you could break your monitor.

      You mean your graphic drivers, right? not your actual hardware?

      (edit: oh no)

      • Truls@mastodon.social
        link
        fedilink
        arrow-up
        4
        ·
        6 months ago

        @andrewth09 I bricked a monitor when I tried to fiddle with the graphics settings in Linux back in the late 90s (tried to get it to run on 1280*1024 - which was considered “hi resolution” back then). I had to buy a new monitor. Then installed Windows and only returned to Linux a long time after that.

  • lordnikon@lemmy.world
    link
    fedilink
    English
    arrow-up
    22
    ·
    6 months ago

    You could buy box copies of the original suse Linux that had manuals in the box the size of a TI graphing calculator manual.

    Once you got X working everything else was cake by comparison.

  • mortalic@lemmy.world
    link
    fedilink
    arrow-up
    19
    ·
    6 months ago

    This was me, you’re talking about me. 😂 In the 90’s Linux was barely getting started but slackware was probably the main distro everyone was focused on. That was the first one I ran across. This was probably late 90’s, I don’t remember when slack first came about though.

    By the time the 2000’s came around, it was basically a normal thing for people in college to have used or at least tried. Linux was in the vernacular, text books had references to it, and the famous lawsuit from SCO v IBM was in full swing. There were distro choices for days, including Gentoo which I spent literally a week getting everything compiled on an old Pentium only for it to not support some of the hardware and refuse to boot.

    There was a company I believe called VA Linux that declared that year to be the year of the Linux desktop. My memory might be faulty on this one.

    Loki gaming was a company that specialized in porting games to Linux, and they did a good job at it but couldn’t make money. I remember being super excited about them and did buy a few games. I was broke too so that was a real splurge for me. I feel like they launched in the 90’s (late) and crashed in the early 2000’s.

    • constantokra@lemmy.one
      link
      fedilink
      arrow-up
      7
      ·
      edit-2
      6 months ago

      I think you need to qualify that having used or tried Linux in college was normal in the 2000s for someone in computer science or engineering, or basically my fellow undiagnosed autistics and autistic adjacents. In my experience it was fairly normal in college for most people to have trouble operating a basic word processor, and they would not have had any idea what Linux was at all.

      • mortalic@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        6 months ago

        Maybe, but I took some business courses too and even some of them had at least tried a Linux distro. I think it was more widespread than just turbo nerds and cs majors. Hell one of the biggest Linux guys I knew was an anthropology major.

  • schizo@forum.uncomfortable.business
    link
    fedilink
    English
    arrow-up
    19
    ·
    6 months ago

    It was prominent in smaller businesses that wanted or needed a Unix but weren’t going to pay what sun or IBM or HP and friends wanted for their hardware+software.

    It ate the proprietary Unix market awfully quickly and I don’t think anyone really misses it.

    For me, educational stuff was all windows with a small amount of macs and I don’t think I ever saw a Linux system in actual use anywhere.

    I used it on the desktop but that was super rare because hardware support was nowhere as good as now - even getting X up was a challenge (go read up on mode lines if you want some entertainment).

    • lemmyvore@feddit.nl
      link
      fedilink
      English
      arrow-up
      6
      ·
      6 months ago

      For me, educational stuff was all windows with a small amount of macs and I don’t think I ever saw a Linux system in actual use anywhere.

      Linux systems started being common in CompSci schools around mid-90s, around the time LAMP took off (fun fact, Apache, MySQL and PHP were all launched in 1995).

      Previously in CompSci you’d get to use all kinds of UNIX servers. My uni still had Solaris servers with dumb terminals, and I got my first sysadmin certification on SCO.UNIX / OpenServer.

      • schizo@forum.uncomfortable.business
        link
        fedilink
        English
        arrow-up
        2
        ·
        6 months ago

        I was doing english lit stuff in that era so what showed up on the tech side was a little different. Ended up spending my entire career actually in IT (25 years now ugh I’m old) because it turns out uh, there’s not any money in a english degree.

  • Shimitar@feddit.it
    link
    fedilink
    arrow-up
    19
    ·
    6 months ago

    Ah, Linux from scratch…

    Also, hardware was… Harder back then, on Linux (mostly modems).

    Beside that, software wise there was less stuff on Linux than today, so you had to check carefully you had what you needed.

    But I was already a Linux user, and a linux-only user at that.

      • cmnybo@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        10
        ·
        6 months ago

        Is was those crappy winmodems that caused all the problems. They cheaped out on hardware, so you basically got a sound card. All of the work had to be done by the driver, which also put a lot of load on your CPU. Serial modems just worked since everything was done in hardware.

        • Quazatron@lemmy.world
          link
          fedilink
          arrow-up
          3
          ·
          6 months ago

          I had erased that information from my memory. Also it took a long time for Linux to gain USB support, then a long time to get WIFI (also because of the cheap vendors that used windows drivers to do the heavy lifting). Yeah, it was a very uphill struggle, with Microsoft actively pushing against Linux (remember the ‘Linux is a virus’ narrative?) I’m amazed we made it this far.

      • Shimitar@feddit.it
        link
        fedilink
        arrow-up
        4
        ·
        6 months ago

        Well, only “real” modems… Those amazing piece of crap that offloaded hardware to the windows driver where… Questionable.

        And they started appearing around windows 95/98.

  • porl@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    ·
    6 months ago

    Hearing your monitor squeal when you got the modelines wrong was fun.

    • gari_9812@lemmy.world
      link
      fedilink
      arrow-up
      5
      ·
      6 months ago

      Could you please elaborate? I’ve no idea what that sentence means, so it sounds really wild to me 😅

      • IsoKiero@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        9
        ·
        6 months ago

        Back when CRT monitors were a thing and all this fancy plug’n’play technology wasn’t around you had modelines on your configuration files which told the system what kind of resolutions and refresh rates your actual hardware could support. And if you put wrong values there your analog and dumb monitor would just try to eat them as is with wildly different results. Most of the time it resulted just in a blank screen but other times the monitor would literally squeal when it attempted to push components well over their limits. And in extreme cases with older monitors it could actually physically break your hardware. And everything was expensive back then.

        Fun times.

      • Aceticon@lemmy.world
        link
        fedilink
        arrow-up
        5
        ·
        6 months ago

        CRT monitors internally use an electron gun which just fires electrons at the phosporous screen (from, the back, obviously, and the whole assembly is one big vacuum chamber with the phosporous screen at the front and the electron gun at the back) using magnets to twist the eletcron stream left/right and up/down.

        In practice the way it was used was to point it to the start of a line were it would start moving to the other side, then after a few clock ticks start sending the line data and then after as many clock ticks as there were points on the line, stop for a few ticks and then swipe it to the start of the next line (and there was a wait period for this too).

        Back in those days, when configuring X you actually configured all this in a text file, low level (literally the clock frequency, total lines, total points per line, empty lines before sending data - top of the screen - and after sending data as well as OFF ticks from start of line before sending data and after sending data) for each resolution you wanted to have.

        All this let you defined your own resolutions and even shift the whole image horizontally or vertically to your hearts content (well, there were limitations on things like the min and max supported clock frequency of the monitor and such). All that freedom also meant that you could exceed the capabilities of the monitor and even break it.

  • SteveDinn@lemmy.ca
    link
    fedilink
    arrow-up
    15
    ·
    edit-2
    6 months ago

    I remember running Slackware and having to recompile the kernel for just about any hardware you added. I configured a box to be used as a router before routers were something you could get commonly at Best Buy.

    I was taking comp. sci. at university and all our work was done on Sparc workstations. Having a Unix-like machine at home was a great help during that time

  • kbal@fedia.io
    link
    fedilink
    arrow-up
    15
    ·
    6 months ago

    My linux experience:

    1993 - Hey, there’s a new Unix-like thing for the PC. You can check it out down at the university computer club.

    1994 - Wow, I finally managed to get X running

    1996 - It was somewhat normal for more nerdy software developers to run linux full-time on their desktop at work.

    1998 - Linux was taking over servers to the point where you rarely saw Solaris, HP-UX, AIX around any more.

    2002 - Everyone agreed that linux was pretty much ready to take over the desktop as well.

  • SeikoAlpinist@slrpnk.net
    link
    fedilink
    arrow-up
    15
    ·
    edit-2
    6 months ago

    It was kind of an upstart thing and people were trying to find ways to monetize it.

    My first Linux was Red Hat on a 486 in 1998 and it was different than I was used to. I was a kid who didn’t know how to startx so I just emailed a developer using pine and they helped me figure out and choose a window manager. Nobody even got mad at this barely teenager just emailing dumb questions. I got lost with fvwm95 and afterstep. I tried every window manager, mlvwm, qvwm, IceWM, etc but ended up liking blackbox the most. I had 12MB of RAM on my first Linux system, 1MB of vram and 256 colors. We were all sarcastic in a cringe, adolescent way but everyone was friendly and helpful.

    There was this fascination with monkeys in pop culture, but not real monkeys --chimps and gorillas. People would throw monkey in their username or in some random nu-metal song for some reason. There were monkeys you could download for your desktop. There was this thing by PC gamer called coconut monkey. I don’t know what that’s all about. And anyway I associate this period with the foot logo of Gnome, which was unprofessional but that was the point. Also, gimp was a funny name for an app (its cringe today), and PAN stood for pimp ass news.

    I discovered Slashdot and Freshmeat and Sourceforge and kuro5hin. Usenet groups were great back then. So was irc. I trolled Slashdot and got negative karma and for the next 15 years before we all moved to SoylentNews, my comments started at -1.

    Nobody knew how to pronounce Linux. Some people said Line-X because his name was Linus like on Charlie Brown, and some people said Leenucks.

    At some point it became a corporate thing and the term Linux was everywhere. Randomly on magazine covers. There was also this divide, almost marketing driven, it seemed that people who liked warez and whatever started to love Microsoft and shit on Linux. So gamers especially started to shit talk and that’s the first time that being a computer nerd wasn’t like this unifying concept, there was an us versus them divide. People who could compile code they wrote and who were genuinely curious versus people who just wanted to download a bunch of shit and show you how big their start menu was and play games. I think this divide still exists.

    There was a bunch of commercial software for Linux too. Metro-X, Accelerated X, Motif, Applixware, Star Office. Descent 3. One of the Quakes. Motif, the toolkit, looked amazing. I thought CDE with themes was the coolest looking thing ever. But I couldn’t afford CDE so I used XFce which was an XForms knockoff. And then enlightenment came along and pushed the boundaries of what we thought a desktop would be. Also, I was able to drag console windows with transparency on that 486 on e16.

    Debian kind of had an elitist community and talked down to people so I never used it. I liked Slackware the most and spent a weekend downloading the floppies over a dialup connection. That led to me discovering FreeBSD in 1999, which I stuck with for almost a decade.

    Later, a comp sci student, I didn’t see Linux at university in the labs. It was Solaris and macOS in the mid 2000s. Eventually, the Solaris computers were shut down and replaced with more Macs.

    My girlfriend’s Windows ME computer was so full of spyware so I installed SuSE with KDE on it for her in her dorm. And she was able to do her papers in AbiWord. And 20+ years later we are married and it all worked out.

    I finally switched to Debian stable about 4 years ago and have no complaints. It’s a lot easier now.

    Edit: A couple more things: I started using Linux because I was very poor and it was free and Windows 95 was a mess on my system. I mean dirt roads and no water for long periods of time. My 486 in 1998 was sort of old already and it came with 8mb of RAM as a hand me down in 1995, but I was dumpster diving outside a community college when I was 12 and found an IBM PS/2 and stole the 30 pin SIMMs out of it. And one of them worked in my 486 computer so I ended up with 12mb of RAM. I overclocked it to 100mhz. That 486 got me through high school and into college where I ended up with an AMD system with a pirated Thai RM233 Windows 2000. But I went back to FreeBSD because I needed a compiler. So that kind of knowledge was useful and now that I have a good career from what I learned, I have donated a lot of money over the years to different projects. Also I make sure my kids have only ever known Linux and Gnome and it works fine for them.

    • InfiniteKrebs@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      6 months ago

      Wow, thanks for sharing all that, was well written and really allowed for a peek into what it was like!

  • Blizzard@lemmy.zip
    link
    fedilink
    English
    arrow-up
    12
    ·
    6 months ago

    Not a veteran, but… During the 90s, while still in primary school, a friend of mine bought a Chip magazine with a CD attached and instructions inside the magazine how to install a mysterious thing called “Linux” from said CD. It was supposed to be something like Windows 95, but new, better and it had a Penguin on it, so we decided to try it.

    We followed magazine’s installation guide to the letter (or at least we thought so) until the installation stuck at error saying KERNEL PANIC!!! and wouldn’t let us finish. We didn’t understand English much back then, but we found the panicking kernel hilarious. Anyway, we figured it’s been enough h4Ck!nG for that day and got back to playing Diablo 1.

  • yak@lmy.brx.io
    link
    fedilink
    English
    arrow-up
    12
    ·
    6 months ago

    If you weren’t at a university it was generally a challenge to get hold of disks. Downloading at home took forever on a 28.8 or even 56k modem (ie. 56 kilobits per second).

    Slackware and Redhat disk sets were the thing, in my experience. But generally that only gave you the compiled code, not the source (although there was an another set of disks with the source packages).

    If you wanted to recompile stuff you had to download the right set of packages, and be prepared to handle version conflicts on your own (with mailing list and usenet support).

    Recompiling the kernel with specific patches for graphics cards, sound cards, modems and other devices (I remember scanners in particular), or specific combinations of hardware was relatively common. “Use the source, Luke!” was a common admonition. Often times specific FAQ pages or howtos would be made available for software packages, including games.

    XFree86 was very powerful on hardware it supported, but was very finnicky. See the other posts about the level of detail that had to be supplied to get combinations of graphics cards and monitors working without the appearance of magic smoke.

    Running Linux was mostly a enthusiast/hobbyist/geek thing, for those who wanted to see what was possible, and those who wanted to tinker with something approaching Unix, and those who wanted to stretch the limits of what their hardware could do.

    Many of those enthusiasts and hobbyists and geeks discovered that Linux could do far more than anyone previously had been prepared to admit or realise. They, and others like them, took it with them into progressively more significant, and valuable projects, and it began to take over the world.

    • constantokra@lemmy.one
      link
      fedilink
      arrow-up
      4
      ·
      6 months ago

      “mailing list and Usenet support”. Yeah. If you’ve ever looked up some weird issues and the only thing that you can come up with is some Debian message group that looks like it was typed on a typewriter, is extremely difficult to follow the response chain, and is apparently from before Y2K… That’s what it was like to run Linux back then.