Start a Conversation

This post is more than 5 years old

Solved!

Go to Solution

11512

May 18th, 2017 02:00

Isilon Vs ECS

Hello Guys,

Isilon can support all modern data center protocols like HTTP , Rest API, HDFS etc,. Along with that it can support SMB and NFS as  NAS storage. Except the remotely anywhere access feature isilon can support all the features where ECS does.

With that all being said, what are the necessity and advantage does ECS contains in par with Isilon from techinical perspective ?

Thanks
Saran

300 Posts

May 18th, 2017 03:00

you are only focussing on the access perspective and not on the backend.

ECS and Isilon are completely different when it Comes to replication.

ECS and Isilon are completely different, when you want to have something "global". The Isilon can scale in a Cluster but if you Need the data "around the world" you will have a lot of Problems with syncs

Isilon does only "1:n" Replication with Write --> Read async

ECS is more softwarebased replication with all write in different datacenters.

It just depends on what you want to do and what you want to achieve.

If you want to host a lot of data in ONE Datacenter and just need the named protocols go with isilon

If you want to host a lot of data in SEVERAL Datacenters and need geo-loadbalancing and active-active-active configurations go with ECS

Furthermore Isilon is more a "Filecentric" solution which offers you access to files through different protocols while ECS is "Objectstorage" which offers you acces to objects through several protocols, this said ECS has WORM by design.

1 Rookie

 • 

108 Posts

May 18th, 2017 04:00

Thanks Sluetze for the explanation.

This one is slightly out of topic, However still to understand better about object storage ,ECS -one of the use case is data analytics , how do they achieve the performance in Big data or other analytics solution with just having the SATA disk bundle ?

300 Posts

May 18th, 2017 05:00

since I have no skills in data anlytics i can't help.

Maybe someone else can jump in?

33 Posts

May 18th, 2017 06:00

In general "big data"/distributed/object protocols like ceph/hadoop/s3 utilize SATA by allowing the reads and writes to scale horizontally with near linear gains.  One SATA disk isn't that impressive but once you get to dozens, hundreds, or thousands you could easily surpass the performance of a high-end storage platform.  The presumption is that the applications have sufficient parallelization as well since a single stream will usually perform no better than what SATA can provide.

Object stores also give up many things POSIX filesystems might take for granted like close-to-open consistency opting for eventual consistency and trade it for strong guarantees that if a file is visible then it's intact.

That said I have not been terribly impressed with the ECS units we have.  They are okay, but not really great.  Frankly I was expecting more given the hardware but it seems like the software platform is just too heavy.  64GB of ram and I often find them swapping due to the nature of their java applications.  Throughput is sufficient for out needs but slow metadata performance ( enumeration ), slow space reclamation after deletion, and we still have issues with larger objects behaving badly.

450 Posts

May 18th, 2017 07:00

Think of ECS in the same way as you may have thought of Centera in the past, a really big archive.  It can do more as sluetze  said with regards to geo-load balancing and so-forth, and supports the Centera CAS API, and S3.  File protocols were added later to ECS initially NFSv3 only, and I think recently SMB as well.  But on the big data side it doesn't speak HDFS like Isilon does, and it's not meant to be best of breed performance for really large HDFS datasets.  As I understand it the HDFS access is actually over NFS using a java library.  So maybe a good spot for cold buckets of data, not hot.

It's really about a different dataset type.  ECS is an object store.  So think of data that has lots of associated metadata like PACS images from radiology where there are tons of identifying DICOM tags, age, date of birth, gender, etc.

Isilon while it can do that do is more about being a traditional file services NAS.  It does HDFS very well because of it's architecture, and is far faster at data loads and much less overhead than DAS attached to data nodes.  Isilon is also really great at big file, giant dataset things like video rendering farms, video surveillance, genome sequencing in life sciences, giant flat file database backups, and then even with some of the more advanced file services features, like snapshots, quotas with notifications, it organically became a good place for large enterprise stores for traditional NAS data, like home directories, group/departmental shares, etc.

 

Just my 2 cents,

~Chris Klosterman

Principal SE, Datadobi

chris.klosterman@datadobi.com

1 Rookie

 • 

108 Posts

May 18th, 2017 21:00

Got it .. Thanks for your reply. So,object storage can't be used for an application with demanding performance.

450 Posts

May 19th, 2017 08:00

Usually no, especially object storage with built-in geo-protection.  One reason why is that most systems like this have data protection policies that require remote protection of the data in some form (either sync or a-sync) before an ack can be sent back of a successful write.  That adds write latency, which could be substantial depending upon the link speed, file size, and latency of the link.

For a read, sure object storage could be quite fast as long as the system you're talking to has a copy of the object in question, and doesn't have to pull it from somewhere else.

No Events found!

Top