My old sandisk SATA ssd was starting to get really slow for some reason. The SMART data and the sandisk SSD dashboard app were saying the SSD was healthy, but its performance wasn’t anywhere near what it was when brand new.
When benchmarked, it was all over the place with looong access times:
Sooo I decided to take the opportunity to upgrade the SSD to something faster - ended up grabbing a Transcend 512GB drive, with onboard DRAM
There were two problems though:
- My motherboard doesn’t support NVMe (at least officially)
- My only available PCIe slot is an x1/single lane
After researching, I realised that the single PCIe lane would still give me almost 1GB/s in real world usage - even though its far from the 3GB/s the drive is rated for, it’s double the speed of SATA and it’s worlds apart from my Sandisk ssd lol.
Ordered an NVMe to PCIe adapter, and proceeded to chop up my PCIe slot to make it fit:
PCMR NSFW
It took a while since I don’t own a dremel 🤪
Once that was done, I kapton taped up the exposed metal bits on the NVMe adapter, that could short on a mobo heatsink nearby.
In it goes!! (The GPU went in after the pic lol)
After re running the benchmarks, OMG the speed difference is insane, although it’s limited by that single PCIe lane.
I was caught off guard by something else though. After cloning my existing install to the new NVMe SSD, it booted right up, with the original Sandisk drive gone. My BIOS does not even recognise the NVMe drive as a disk drive, and there are no settings anywhere in there for it.
BIOS person, thank you whoever you are, you saved me needing to do more jank to get my unsupported NVMe drive working!
I am more than happy so far with the dramatic speed increase compared to the SATA drive. I can now actually shut down my desktop when I’m not using it 🥲
TIL you can install PCI cards by just…taping up extraneous pins??? And they work that way??? Cooooool.
PCIe is extremely flexible. Often times those big 16x slots are actually only wired for 8x or even 4x. The device has to be able to support running on fewer lanes than expected because of stuff like this.
Well, its basically running at PCIe 1x. Thats why theres all those notches on the full 16x card. You can install in a 1x, 2x, or 4x header, and it will work. Problem is some cards dont have the notches and need the full connector, but supposedly you can grind off some plastic from the port to make it fit and it will still work.
supposedly you can grind off some plastic from the port to make it fit and it will still work.
You absolutely can do this, often slots for 1x or 4x slots come with open backs for this reason - it’s better than only wiring up that many lanes in a 16x slot IMO!
What gets me is the number of people who take a saw to a pci-e card instead of just opening the back of the slot up… :(
Also this is PCIe, not PCI. It’s just got a few gaps because it’s actually only x4, and the other pins are just there for stability.
You don’t even need to tape them up. The slot is open ended explicitly for this kind of situation. The extra contacts just dangle and have no power running through them since they aren’t connected.
Ive never actually seen an open-ended slot though, and OP had to use a Dremel to open his. Probably for the exact reason that OP added the kapton tape, because it would be stupidly easy to short out those exposed pins.
Oh? Weird I thought it was standard. Looking at the PCs in my house it seems I was very wrong, none of them have an open ended 1x.
Maybe it’s the mandela effect.
Some of my server boards do, so they do exist, but I haven’t seen it much on consumer stuff.
deleted by creator
Yeah, this is where I ended up as well. I could have sworn I currently owned a few boards with it, but of the 7 I checked none of them had it.
deleted by creator
I’ve seen some, but they’re less common now and you’d only see 1 or 2 on a board.
Now they mostly out larger slots but don’t connect the pins.
This is cursed and I love it!
Cutting the back of x4/x8 slots is pretty common. Some motherboards even have open back slots as default.
Great work, OP. A very nice and well thought out upgrade (most folk skip the kapton tape).It’s worth everyone knowing that pcie slots are generally universally compatible. So a GPU (wanting 16x slot) can run in a 1x physical slot (infact this is how GPU mining expansion rigs work).
I don’t think I’ve ever encountered a pcie card that will not function at all unless it receives it’s required lanes.I don’t think I’ve ever encountered a pcie card that will not function at all unless it receives it’s required lanes
One of the few things that’d be problematic would be the x16 -> quad M.2 cards which use PCIe bifurcation.
Lanes 1-4 from the socket are wired to the first M.2, 5-8 to the 2nd, etc.It would still work (by some definition of the word), but in the sense that the first M.2 drive would get 1 lane and any others wouldn’t be connected.
(Quad M.2 boards with a “PLX” or other PCIe switch chip would work fine with 1 upstream lane serving all 4 drives)
the fact that it fires up without BIOS recognition is cool and weird… that moment it boots after that kind of migration upgrade is like nothing else…
and i gotta say awesome mod dude… that was one cool ass solution with available hardware…
If you put your bootloader with nvme support on a recognized drive it works.
thank you that makes perfect sense, and is pretty slick
How much space was left on the original 256GB disk? A common theme i heard about solid state drives was that speed will suffer the more space is used, so idealy you should keep it ~50% free for optimal performance.
deleted by creator
The original ssd was rammed full, only about 20GB free. Writing files was super fast - it was mainly reading files that the SSD struggled to do at a palatable speed, often locking up my pc for a few seconds
With the new SSD i’ve just left the remaining space unallocated, so it’s firmware can freely use that space for whatever it needs to 😁
While this does depend on the drive… you really never want to get above 85-90% full on an SSD.
It really does depend on how the drive deals with partially filled blocks, caching, and over provisioning…
But I’ve seen this massive slow down worse than spinning disk on some older Samsung consumer grade drives when full. Now I just shrink the usable space by 10% on a new SSD and don’t worry about it.
https://pureinfotech.com/why-solid-state-drive-ssd-performance-slows-down/
Too bad we don’t have a similar community to TechSupportMcGyver
I’m confused, why didn’t you get a 1x pcie carrier board?
Price mainly, and I wanted something that wouldn’t be constrained by a 1x lane in another mobo in future
This is some bde hacktastic shit and I’m all here for it.
Great job.
Love seeing a dodgy project come together and deliver real world results 💪
So long as it’s working that’s all these matters
…so that’s what the PCIe x1 slot for, huh?
Ive got a few 1x cards. For example, my old home server box has a 1x slot that you can plug in a 1x gpu, or additional networking card, USB, etc.
Lol. You can shove all sorts of crap in those.
Nice job! This is the type of stuff I come here for.
That’s a really neat upgrade :) props. I’m guessing you have some additional drives with the bootloader on them. While the bios may not support nvme as a boot device, as long as the boot loader is on a supported drive and supports nvme drives it can boot the nvme drive. Had to do this once with a flash drive.
Aren’t you still bound by the limits of the SATA standard? So while the NVMe might have cheaper storage, the speed is bottle necked.
No. The back end interface is pcie instead of sata.
Is NVMe really cheaper? I thought SATA was the cheaper one.
NVME has been cheaper for a while now.
handy website for comparing disk prices
Props to you, awesome job.