Comments Locked

58 Comments

Back to Article

  • "Bullwinkle J Moose" - Thursday, November 9, 2017 - link

    Humor me.....

    How fast can you copy and paste a 100GB file from and to the same Optane SSD

    I don't believe your mixed mode results adequately demonstrate the internal throughput

    At least not until you demonstrate a direct comparison
  • Billy Tallis - Thursday, November 9, 2017 - link

    Your concept of "internal throughput" has no basis in reality. File copies (on a filesystem that does not do copy-on-write) require the file data to be read from the SSD into system DRAM, then written back to the SSD. There are no "copy" commands in the NVMe command set.
  • "Bullwinkle J Moose" - Thursday, November 9, 2017 - link

    "There are no "copy" commands in the NVMe command set."
    ---------------------------------------------------------------------------------
    That might be fixed with a few more onboard processors in the future but does not answer my question

    How fast can you copy/paste 100GB on THAT specific drive?
  • "Bullwinkle J Moose" - Thursday, November 9, 2017 - link

    Better yet, I'd like you to GUESS how fast it can copy and paste based on your mixed mode analysis and then go measure it
  • Lord of the Bored - Friday, November 10, 2017 - link

    How will a new processor change that there is no way to tell the drive to do what you want? We don't trust storage devices to "do what I mean", because the cost of a mistake is too high. No device anyone should be using will say "it looks like they're writing back the data they just read in, I'mma ignore the input and duplicate it from the cache to save time." Especially since they can't know if the data is changed in advance.

    Barring a new interface standard, it will take exactly as long to copy a file to another location on the same drive as it will to read the file and then write the file, because that is the only provision within the NVMe command set.
  • "Bullwinkle J Moose" - Thursday, November 9, 2017 - link

    What would happen if Intel Colludes with AMD to implement this technology into onboard graphics instead of AMD's plan to use Flash in their graphics cards ?

    Seems to me like Internal throughput would be very important to the design
  • Samus - Thursday, November 9, 2017 - link

    That is also file system dependent. For example, in Mac OS High Sierra, you can copy and paste (duplicate) any size file instantly on any drive formatted with APFS.

    But your question of a block by block transfer of a file internally for a 100GB file would likely take 50 seconds if not factoring in file system efficiency.
  • cygnus1 - Thursday, November 9, 2017 - link

    That's not a copy of the file though. It's just a duplicate file entry referencing the same blocks. That and things like snapshots are possible thanks to the copy on write nature of that file system. But, if any of those blocks were to become corrupted, both 'copies' of the file are corrupt.
  • "Bullwinkle J Moose" - Thursday, November 9, 2017 - link

    Good call Samus

    I noticed that the Windows Fall Crappier Edition takes MUCH longer to copy/move/paste in several tests than the earlier versions of "Spyware Platform 10"

    as well as gives me a "Format Disk Now?" Prompt after formatting a new disk with Disk Manager
    as well as making a zero byte backup with Acronis 2015 (Possible anomaly, Will test again)
    as well as breaking compatibility with several programs that worked in earlier versions
    as well as asking permission to do anything with my files and then often failing to do it
    as well as, well.....you get the idea, they fix one thing and break 5 more

    Disclaimer:
    I do NOT believe that xpoint could be used in its current form for onboard graphics!
    But I'd like to know that the numbers you are getting at AnandTech match your/my expectations and if not, why not?

    Sorry if I'm sounding like an AHole but I'd like to know what this drive can really do, and then what Microsoft graciously allows it to do ?

    make sense?
  • PeachNCream - Friday, November 10, 2017 - link

    It doesn't make sense to even care what a client OS would do with this drive. It's enterprise hardware and is priced/positioned accordingly. Instead of the P4800X, check out the consumer 900p version of Octane instead. It's a lot more likely that model will end up sitting in a consumer PC or an office workstation.
  • AlaskanDruid - Thursday, November 9, 2017 - link

    Ok. what am I missing? What is the use of these when you can have one in a m.2 form which takes up MUCH less room?

    I know I'm missing something, but what is it?
  • Drumsticks - Thursday, November 9, 2017 - link

    The drives dissipate 12-18W, which is 2-3x more than flash based SSDs. I don't think the m.2 form factor would make heat dissipation easy, unless you scale the design back down to the 32-120GB range.
  • Billy Tallis - Thursday, November 9, 2017 - link

    More than just capacity, you'd almost certainly have to scale down performance by a lot. It's possible to do better than the current PCIe x2 32GB module, but a 7 channel controller would be hard to fit physically, and it would have to run slower to not melt. I think a ~3 channel M.2 drive is plausible, but I'm not sure its write speeds would be high enough.
  • DanNeely - Thursday, November 9, 2017 - link

    it looks like you could physically do 7 (or even 8) channels on a dual sided 22110 m.2 module with the current package size. The thermals would be a real challenge though; and are why I really wish U.2 would've gained at least a little traction in the consumer market.
  • Billy Tallis - Thursday, November 9, 2017 - link

    I was thinking more along the lines of a 22x80mm card. I agree that 7 3D XPoint packages is probably possible on 22x110, but I'm not sure how dense the controller's BGA pad can get.
  • DanNeely - Thursday, November 9, 2017 - link

    If you went 2 sided you could fit 6 on a 2280 card. With a bare bottom they're just short on enough room for a 3rd package on top; move some of that misc stuff underneath and you'd free up the PCB area needed.
  • Rwbeckman - Thursday, November 9, 2017 - link

    This is the datacenter model, good for releasing higher capacity sooner with its size. Plus rackmount cooling design will help. U.2. Will be for hot swap, the card will not. The M.2 enthusiast consumer model is the 900P
  • Samus - Thursday, November 9, 2017 - link

    XPoint technology is just one of the reasons Intel released the "ruler" form factor for future SSD's. I suspect a mobile variant will replace m2 in the future.

    https://www.anandtech.com/show/11702/intel-introdu...
  • CheapSushi - Thursday, November 9, 2017 - link

    Or just use the new M.3 format (wider) that Samsung came out with.
  • Drumsticks - Thursday, November 9, 2017 - link

    Great review Billy, thank you!

    Optane as a gen 1 product is pretty fantastic. I can't wait to see how the rest of the market will compete, and what Intel and Micron (where are their drives?) can do for round two.
  • Elstar - Thursday, November 9, 2017 - link

    I thought the whole point of 3D XPoint memory would be that it is DIMM friendly too.
    1) When will we see this in DIMM form?
    2) Would the DIMM version also need/have the reserved/spare capacity?
    3) Why is this spare capacity even needed? 1/6th seems like a potentially high internal failure rate (or other fundamental problem.)
  • PeachNCream - Thursday, November 9, 2017 - link

    It seems like current 3D XPoint doesn't have enough endurance yet to sit in a DIMM slot unless it's gonna be just a storage drive in DIMM form factor. That and because we're only just now seeing early enterprise and retail products, I bet that we're gonna need another generation or two before we get DIMM versions. :(
  • Billy Tallis - Thursday, November 9, 2017 - link

    Intel hasn't said much about 3D XPoint DIMMs this year, other than to promise we'll hear more next year.

    It's not clear how much fault tolerance Intel is trying to build in to the Optane SSDs with the extra capacity. A bit of it is necessary to support sector formats with protection information (eg. 16 extra bytes per 512B sector, or 128B extra for each 4kB sector). Beyond that, there needs to be room for the drive's internal data structures, which aren't as complicated as a flash translation layer but still impose some space overhead. The rest is probably taken by a fairly simple erasure coding/ECC scheme, because it's almost impossible to do LDPC at the speed necessary for this drive. (That's also why DIMMs use simple ECC instead of more space-efficient codes.)
  • woggs - Thursday, November 9, 2017 - link

    Most all intel SSDs have a parity die, so one full die likely provides internal raid protection of data. The rest is for ECC, internal system information, media management and mapping out defects... Impossible to know which of these is driving the actual spare implemented. I count 14 packages, so 1/14th (7%) is already the internal parity. 16% is big relative to nand consumer SSDs but comparable to enterprise. Doesn't seem particularly out of line or indicative of something wrong.
  • CheapSushi - Thursday, November 9, 2017 - link

    Micron is probably working on the DIMM version.
  • woggs - Tuesday, November 14, 2017 - link

    Intel is working on DIMMs... "Now, of course, SSDs are important, but in the long run, Intel also wants to have Optane 3D XPoint memory slot into the same sockets as DDR4 main memory, and Krzanich brought a mechanical model of an Optane DIMM to show off." https://www.nextplatform.com/2015/10/28/intel-show...
  • MajGenRelativity - Thursday, November 9, 2017 - link

    I enjoyed the review. Keep up the good work!
  • melgross - Thursday, November 9, 2017 - link

    I’m curious as to how this will perform when PCI 4 is out next year. That is, one with a PCI 4 interface. How throughput limiting is PCI 3 for this right now?
  • MajGenRelativity - Thursday, November 9, 2017 - link

    It shouldn't be that limiting, as PCIe 3.0 x4 allows for a higher throughput than 2.4 GB/s. There could be some latency improvements (probably small), but I don't think throughput is the issue
  • woggs - Thursday, November 9, 2017 - link

    If the question is "could a drive be made to saturate gen 4?" then, yes, of course, if intel chooses to do so. Will require a whole drive. Latency is a more interesting question because that is what 3dxp is really providing. QD1 latency is <10us ((impressive!). I don't expect that to improve since it should be limited by the 3dxp itself. The PCIe and driver overhead is probably 5us of that. Maybe gen 4 will improve that part.
  • woggs - Thursday, November 9, 2017 - link

    This drive is PCIe 3, so nothing will change because it will link as gen 3. Will need a whole new drive...
  • MajGenRelativity - Thursday, November 9, 2017 - link

    I believe OP was talking about a drive that was identical to this in every way, but supporting PCIe 4
  • woggs - Thursday, November 9, 2017 - link

    And SSD is a system and would neither over nor under engineer any one piece significantly. So, altering one piece should not be expected to suddenly pop to some much higher performance. Will require a whole new drive.
  • Hurr Durr - Thursday, November 9, 2017 - link

    Where is requisite ddriver hysterics session about hypetane?
  • Reflex - Thursday, November 9, 2017 - link

    It appears the hysteric is gone at least for the time being. As a result the discussion has been far easier to read.

    Great drive, can't wait to see future generations where presumably they will increase density and reduce power consumption...
  • woggs - Thursday, November 9, 2017 - link

    Yep!
  • woggs - Thursday, November 9, 2017 - link

    (on both counts)
  • Hixbot - Thursday, November 9, 2017 - link

    He must be taking the day off. I'm sure he'll flood the comments tomorrow.
  • Makaveli - Thursday, November 9, 2017 - link

    If you say his name 3 times he will show up....

    DON'T DO IT!
  • abhaxus - Thursday, November 9, 2017 - link

    I came here looking for him, and was disappointed.
  • Lord of the Bored - Thursday, November 9, 2017 - link

    Me too. ddriver is most of why I read the comments.
  • mkaibear - Friday, November 10, 2017 - link

    He is always good for a giggle. I suppose he's busy directing hard drive manufacturers to make special hard drive platters for him solely out of hand-gathered sand from the Sahara. Or something.

    Still it's a shame to miss the laughs. It's always the second thing I do on SSD articles - first read the conclusion, then go and see what deedee has said. Ah well.
  • extide - Friday, November 10, 2017 - link

    Please.. don't jinx us!
  • rocky12345 - Thursday, November 9, 2017 - link

    Interesting drive to say the least. Also a well written review thanks.
  • PeachNCream - Thursday, November 9, 2017 - link

    30 DWPD over the course of 5 years turns into a really large amount of data when you're talking about 750GB of capacity. Isn't the typical endurance rating more like 0.3 DPWD for enterprise solid state?

    So this thing about Optane on DIMMs is really interesting to me. Is the plan for it to replace RAM and storage all at once or to act as a cache or some sort between faster DRAM and conventional solid state? Even with the endurance its offering right now, it seems like it would need to be more durable still for it to replace RAM.

    Oh (sorry case of shinies) could this be like a DIMM behind HBM on the CPU package where HBM does more of the write heavy stuff and then Optane lives between HBM and SSD or HDD storage? Has Intel let much out of the bag about this sorta thing?
  • Billy Tallis - Thursday, November 9, 2017 - link

    Enterprise SSDs are usually sorted into two or three endurance tiers. Drives meant for mostly-read workloads typically have endurance ratings of 0.3 DWPD. High-endurance drives for write-intensive uses are usually 10, 25 or 30 DWPD, but the ratings of high-endurance drives have decayed somewhat in recent years as the market realized few applications really need that much endurance.
  • lazarpandar - Thursday, November 9, 2017 - link

    Can this be used to supplement addressable system memory? I remember Intel talking about that during the product launch.
  • Billy Tallis - Thursday, November 9, 2017 - link

    Yes. It makes for a great swap device, especially with a recent Linux kernel. Alternatively, Intel will sell it bundled with a hypervisor that presents the guest OS with a pool of memory equal in size to the system's DRAM plus about 85% of the Optane drive's capacity. The hypervisor manages memory placement, so from the guest OS's perspective the memory is a homogeneous pool, not x GB of DRAM and y GB of Optane.
  • tuxRoller - Friday, November 10, 2017 - link

    It's a bit odd Intel would go for the hypervisor solution since the kernel can handle tiered pmem and it's in a better position to know where to place data.
    I suppose it's useful because it's cross-platform?
  • xype - Friday, November 10, 2017 - link

    I’d guess a hypervisor solution would also allow any critical fixes to be propagated faster/easier than having to go through a 3rd party (kernel) provider?
  • tuxRoller - Friday, November 10, 2017 - link

    Since this is for enterprise, the os vendor would be the one responsible (so, yes, third party) and one of the reasons why you pay them ridiculous support fees is for them to be your single point of contact for most issues.
  • tuxRoller - Friday, November 10, 2017 - link

    Very nice write-up.
    Might it be possible for us to get an idea of the difference in cell access times by running a couple tests on a loop device, and, even better, purely dram-based storage accessed over pcie?
  • Pork@III - Friday, November 10, 2017 - link

    Has no normal only speed test? What are these creepy creations of this vc that?
  • romrunning - Friday, November 10, 2017 - link

    Is there any tests of the 4800X in a virtual host? Either Hyper-V or ESX, running multiple server OS clients with a variety of workloads. With the kind of low latency shown, I'd love to see how much more responsive Optane is compared to all flash storage like a P3608. Sort of a" rising tide floats all ships" kind of improvement, I hope.
  • Klimax - Sunday, November 12, 2017 - link

    That's nice review. How about some test using Windows too. (Aka something with more advanced I/O subsystem)
  • Billy Tallis - Monday, November 13, 2017 - link

    I'm not sure what you mean. Nobody seriously considers the Windows I/O system to be more advanced than what Linux provides. Even Intel's documentation states that the best latency they can get out of the Optane SSD on Windows is a few microseconds slower than on the Linux NVMe driver, and on Linux a few more microseconds can be saved using SPDK.
  • tuxRoller - Tuesday, November 14, 2017 - link

    "Advanced" may be the wrong way to look at it because ntkrnl can perform both sync and async operations, while Linux is essentially a sync-based kernel (the limitations surrounding its aio system are legendary). However, by focusing on doing that one thing well the block subsystem has become highly optimized for enterprise workloads.
    Btw, is there any chance you could run that block system (and nvme protocol, if possible) overhead test i asked about?

Log in

Don't have an account? Sign up now