The maintenance bill is due for your on-premises SAN/NAS–or it just increased. It’s hundreds of thousands or millions of dollars just to keep your existing storage gear under maintenance. And you know you will need to purchase more storage capacity for this aged hardware. Do you renew and commit another 3-5 years by paying the storage bill and further commit to a data center architecture? Do you make a forklift upgrade and buy new SAN/NAS gear or move to hyperconverged infrastructure? Do you move to the AWS cloud for greater flexibility and agility? Will you give up security and data protection?
The following is a recording and full transcript from a webinar that aired live on 08/30/16. You can download the full slide deck on Slideshare.
Full Transcript: AWS vs. On-Premises NAS Storage
Taran Soodan: Good afternoon everyone. My name is Taran Soodan, and welcome to our webinar today on on-premises NAS upgrade, paid maintenance or public cloud. Our speaker today is going to be Kevin Brown who is a solutions architect here at SoftNAS. Kevin, do you want to go ahead and give a quick hello?
Kevin Brown: Hello? How are you guys doing?
Taran: Thanks, Kevin. Before we begin the webinar, we just want to cover a couple of housekeeping items with you all. In order to listen to today’s webinar, you’ll click on the telephone icon that you see here on the GoTo Meeting Control Panel.
Any questions that you might have during the webinar can be posted in the questions box, and we’re going to be answering those questions at the end of the webinar so please feel free to ask your questions.
Finally, this webinar is also being recorded. For those of you who are unable to make it or have colleagues that this might be of interest to, you’ll be able to share the webinar according with them later on today. Please keep an eye out for your email and we’ll send that information your way.
Also, as a bonus for attending today’s webinar, we are going to be handing out $100 AWS credits, and we’ll have more information about that later on at the end of the webinar.
Finally, for our agenda today, we’re going to be talking about on-premises to cloud conversation. We’re going to show you the difference between on-premises versus hyperconverged versus AWS.
We are going to demo how to move your on-premises storage to AWS without having to modify any of your applications. We will also tell you here is why you should choose AWS over on-premises and hyperconverged.
We’ll also give you some information about the actual total cost for ownership for AWS, and we’ll show you how to use the AWS TCO Calculator. Then we will tell you a little bit about SoftNAS and how it helps with your cloud migrations.
Finally, we’ll have a Q&A at the end where we answer any and all questions that you might ask. With that being said, I’ll go ahead and hand it over to Kevin to begin the webinar.
Kevin, you are now the presenter. Thanks, Kevin. I think you might also be on.
Kevin: Not a problem. Can you guys see my screen?
Taran: Yes, we can.
Kevin: All right, perfect. Good morning, good afternoon, goodnight, and we appreciate that you’ve logged in today from wherever you are in the world. We thank you for taking the time and joining us for this webinar.
Today, we’re actually going to focus on a storage dilemma that we have happening with IT teams and organizations all across the world. The looming question is what do I do when my maintenance renewal comes up?
Teams are left with three options. Either we stay on-premise and pay the renewal fee for your maintenance bill which is a continuously increasing expense, or you could consider a forklift upgrade where you’re buying a new NAS or SAN and moving to a hyper-converged platform.
The drawback with this option is that you still haven’t solved all your problems because you’re still on-prem, you’re still using hardware and the new maintenance renewal is about 24 to 12 months away.
Finally, customers can Lift and Shift their data to AWS – hardware will no longer be required, and data centers could be unplugged. Does this sound familiar to anybody on this call?
There is an increase in maintenance costs for support, for swapping disks, for downtime. There is exorbitant pricing for SSD drive and SSD storage when you paying a nominal leg for pricing that you need to pay to ensure that your environment works as advertised.
You have a never-ending pressure from business to add more storage capacity – we need it, we need it now, and we need more of it. There is a lack of low-cost high-performance object storage, and you’re pressured by the business owners for agile infrastructure.
The business is growing; data is growing. You need to be ahead of the curve and way ahead of the curve to actually keep up. Let’s take a look. Let’s do a stare and compare of all these three options that are there so On-Premise, Hyper-Converged, and AWS Cloud.
From a security standpoint, all these three options deliver a secure environment with all the rules and policies that you’ve already designed to protect your environment — they travel with you.
From an infrastructure and management scenario, your on-premise and hyper-converged still requires your current staff to maintain and update the underlying infrastructure.
That’s where we’re talking about your disk swaps, your networking, your racking and un-racking. AWS can help you limit this IT burden with its Manage Infrastructure.
From a scalability standpoint, I dare that you call your NAS or SAN provider and tell them that you think you bought too much storage last year and you want to give them back some. Tell me how that plays out.
I say this in jest but, in AWS, you get just that option. You have the ability to scale up or scale down allowing you to grow as needed and not as estimated. We talked a bit, on the last slide, about the infrastructure and how AWS can help lessen some of that IT burden.
For your on-premise and hyper-converged system, you control and manage everything from layer one all the way up to layer seven. In an AWS model, you can remove the need for jobs like maintenance, disk swapping, monitoring the health of your system and hand that over to AWS.
You’re still in control of managing user-accounts access in your application, but you could wave goodbye to hardware, maintenance fees in your forklift upgrades.
With that, I’d like to share eight reasons for you to choose AWS for your storage infrastructure. From a scalability standpoint. We talked about this earlier. Giving you the ability to grow your storage as needed, not as estimated. For the people who are storage gurus, you know exactly what that means.
I’ve definitely been in rooms sitting with people with predictive modelling about how much data we are going to grow by for the next quarter or for the next year. I could tell you for 100% fact, I have never ever been in a room where we’ve come up with an accurate number. It’s always been an idea, a hope, a dream, a guess.
With AWS and the scalability that it provides, we give you the ability to grow your storage and then pay for it as you go and only pay for what you use. That in itself is worth its own weight in gold.
Not only that, you get a chance to end your maintenance renewals and no longer pay for that maintenance ransom where they are holding access to your data until your maintenance ransom is paid.
There is also huge benefits for trading-in CAPEX for the OPEX model. There is no more long-term commitment. When you’re finished with the resource, you send it back to the provider. You’re done with using your S3 disk, you turn it off and you send that back to Amazon.
You also gain freedom from having to make a significant long-term investment for equipment that you and I know will eventually breakdown or become outdated. You also have a reliable infrastructure.
We’re talking about S3 and its (11) nines worth of durability. EC2, where it gives you (4) nines worth of accessibility for your multi-AZ deployments. You have functions like disaster recovery, easy of use, ease to manage, and your utilizing the best in the class in security to protect your data.
Taran: Thanks for that, Kevin. What we are going to do now is just ask a quick poll question to the audience here. If you’re currently using an on-premises NAS system and it’s coming up for maintenance renewal, what do you intend to do?
Are you going to do an in-place upgrade or you use your existing hardware, but you update the software? Are you going to go a forklift update where you buy new hardware and software?
Or are you going to move to a hyper-converged system? Or are you considering the public cloud whether it’s AWS or other options? I’ll go ahead and give just a few more seconds for everyone to answer.
Looks like about half of you have voted already, so I’ll give that about 10 more seconds and I’ll close the poll. I’ll go ahead and close the poll and share the results with everyone.
As you can see here, we have a good mix of what you all intend to do. Most of you are intending to move to the public cloud, whether it’s AWS or others. Looks like a lot of you are interested in in-place NAS and SAN upgrade, so it’s interesting.
A lot of you are also considering moving to hybrid-converged. For those of you who answered other, we would be curious to learn more about what your other plans are. On the questions pane, you’re more than welcome to write what you’re intending to do.
With that, I’ll go ahead and hand it back to Kevin to handle the demo. Kevin.
Kevin: Not a problem. Can you guys see my screen?
Taran: I can.
Kevin: Perfect. For the sake of my feed, I went ahead and did a recording of this just to make sure that no issues or problems would happen with the demo, so I can walk through. I’m going to talk through this video that I’m playing.
We’re talking about lifting and shifting to the cloud and that can be done in multiple ways. We’re talking about, from a petabyte scale, you can incorporate AWS Import using Snowball. You can connect directly using AWS Direct Connect, or you could use some open-source tools or programs like Rsync, Lsync, and Robocopy.
Once your data is in the cloud, it’s how you’re going to maintain that same enterprise-level of experience that you’re used to. With SoftNAS, you have that ability. We give you the opportunity to be able to have no downtime SLA and we’re the only company that gives that guarantee.
We will be able to walk through the demo. Let me show you. One of the things I’m going to show is the easy use that we have in what we actually do with attaching storage and sharing that storage out to your data consumers.
In this demo, we’re going to go ahead and we’re going to add a device. We are going to add some S3 storage. Very simple and easy, we click “add device.” It comes up “Cloud Disk Expander,” and we’re going to choose Amazon Web Services, S3.
We’re going to go ahead and we’re going to click next. Because we’ve defined our IAM Role, you don’t have to share your access keys or security keys. We do give you the ability to go in and select your region.
For this demo, my instances exist in Northern California so I’m going to select Norther California. If you had an S3 region that’s closer to you, you’d have the ability to do that also.
We give you the ability to choose the bucket name and we increment up the S3 buckets as they are actually created. For the sake of this demo, I’m going to go ahead. I’m going to select a 10GB drive just to make sure that this creates quickly and easy for you to see.
We also give you the ability to encrypt that disk – that would be Amazon’s encryption that we allow you to bleed through. We give you the ability to also encrypt data at a different level.
From this, we see that our S3 disk has been created and is now available to be assigned to a storage group. We are also going to add some EBS disk. We know S3 is not Amazon’s most performing disk. For your data that requires more performance backend, we allow you to add Amazon’s EBS disk.
With that, we already told you IAM there so we don’t have to use our key. For the sake of this, we’re going to do 15GB. We have the ability to encrypt it again. This is disk level encryption.
Then we also give you the ability to choose the storage that is best suited towards your purpose. We have General Purpose SSD. You could use Provisioned IOP disks for your more performance data, or you could choose to use your standard.
For the sake of this demo, we will choose general purpose and I will select to use three. With that, we’re just going to go and we’re going to create that EBS disk.
As you see, the wizard in the background is creating it. It’s going through the process of creating, attaching, and partitioning those disks to be used by the SoftNAS device.
We give it a couple of seconds. We’ll see that this completes. All right, now we have our system. We have that both our S3 and our EBS disk, they have the ability for us to assign them to our storage pool.
Our storage pool is what we use to aggregate our disk. We have then the ability to add some enhancements while we go through that process. We’ll go ahead and we’ll select “Create.”
At this point in time, I am very descriptive with what I name my disk. I am going to call this EBS pool. I’m going to go ahead and select RAID zero and that’s because we’re already working highly-performing redundant disk.
At that point, nothing that I add can add to the performance system that Amazon has, but we can stripe across those disks, which should add performance going in. we’ll go ahead.
We’ll select that drive, and we’ll go ahead. If we go ahead, it’s coming up saying that we’ve choose no RAID and that’s okay because we’re talking about cloud and highly redundant disks. We’re in.
We’re going to do the same thing for our S3 pool. We’re going to go in. We’re going to call it S3_pool if I could spell it correctly. We’re going to go ahead and we’re going to select our S3 storage. We’re going to select RAID zero so we could stripe across.
Remember, stripping across is similar to what you would have in your environment with a server. The more disks you’re stripping across, the more performance you make in a system.
We’re going to go ahead. We’re going to create this pool. Now we have an EBS pool and we also have an S3 pool set up. We’re going to go in. These are some of the other enhancements that we could add to the system.
In front of my S3 pool, I’m going to add some read cache. With the system that I chose, I have ephemeral drive associated with it. This is m3.xlarge. With that, I’ll be able to put that ephemeral drive in front of my system as a way to make my S3 disk be more performing.
I’m going to go ahead. I’m going to select it. Just as simple as that and select it in “Select/Add Cache.” Just as simple as that, I’ve made my S3 disk more performing as I’m utilizing not only striping, but I’m also utilizing the fact that now “read cache” is in front of that system to make those disks more performing.
Now that we’ve added our disk devices, we’ve aggregated them in a storage pool. The next thing that we have to do is that we have to share this out to our end-users, to our applications, to our servers.
The way that we do that is we create volumes in nodes. The way that this is going to share out, your volume name should adhere to your company policies. Whatever share you’re used to sharing out, you should adhere to those rules.
From a volume name standpoint, we’re going to go with user-share, and I’m going to go ahead and select that storage pool. It’s S3. I’m going to select the storage pool.
I don’t need my user-shares and access to my user-shares to be the most performing. I’m going to share it out via NFS and via CIFS. I also have the ability to make a thin provisioned or a thick provisioned.
We also have the ability to do compression and deduplication at this level. They both come with a resource cost. If you are looking at doing compression, you’re going to add a little bit more CPU. For dedupe, you’re going to want to add some more memory.
We’ll talk about snapshots. We get right out of the box. By default, we enable snapshotting of the system and that gives us the ability to be able to sustain the readable/writable clones which we will show you later.
By default, we snapshot the volume every three hours, but it’s definitely tunable if you need something else that would be better for your environment, your LAN, or you could schedule it for something.
We also have a scheduled snapshot retention policy. We start of ten-hourly, six-daily, one-weekly. We could go ahead, and we could create that volume. Now we have a volume created that allows you to access it within a cloud via CIFS or via NFS and it’s backed by S3 storage.
We now are going to go ahead and create a more performing storage. I’m setting up a website that I’m going to have in the cloud, and I want to be able to have some fast disk that’s sitting behind it.
I probably wouldn’t call this a website in production. However, for the sake of this demo, I’m going to go ahead and call it website. I go ahead. I select my storage pool. It’s an EBS disk.
I’m going to go ahead. I’m going to share this out via NFS, and I’m sharing it out via CIFS. We talked about compression and duplication and snapshots. We’re going to go ahead, and we’re just going to do create.
Now we have both volumes created; they are backed by AWS disk. We’re going to go in and we’re going to talk about snapshotting.
QA did something that they ended up breaking our environment. Based on the fact that they broke our environment, they want to be able to test that the fix that they have created works.
What I’m going to do is I’m going to go in. I can’t let them use my production data. What I could do is do a snapshot of my existing environment, and I’m going to create a readable/writable clone so that they could test their fix on that. It’s just as simple.
This is now a point in time instance of the data that existed in website and it’s useable, readable, writable, and shared out in my environment to my users. We also have the ability to attach and integrate with LDAP and Active Directory to ensure that security of your environment is intact.
With that, that is in a nutshell what we went about doing for giving you access to your data. I’m going to run another video. This video should be a little faster. We’re going to go in, and what we’re actually going to do is set up snap replicate between two instances within AWS.
We talked about the SLA within Amazon that gives you five nines worth of availability for instances that you have in multiple AZs. SoftNAS allows you to do that.
If have two instances in this environment, both are in the west. One is in west IA and one is in west IB. with that, I am going to set up replication of my data between both of my instances.
The first thing that we’re going to go in is that we require that the storage size be the same in both nodes. It doesn’t have to be the exact same type of storage, but it needs to be the same amount of storage.
We also require that you name the storage pools the same. I’ll go in. For the sake of time, I went ahead and I added the drives in advance. With that, I’ll just go ahead and I’ll create my storage pools.
Let me create my EBS storage pool if I could spell it correctly, again. Selecting RAID zero, and then I’m going to go ahead and select “Create.” We’re being warned again.
We are very concerned about your data. We want to make sure that it’s extremely secure. With that, we have EBS pool. We’re going to go ahead and we’re going to create our S3 pool or pool of S3 disks.
We’re going to go and we’ll be able to create. Now we have both systems ready to be able to do replication between them. Everything that exists on the lines and volumes, those will be automatically synched from the source node over to the target.
Just verifying one last time to make sure that the pool names are correct in there. Now we should be able to go to snap replicate and set up the snap replication between the source and the node.
We’re going to go in. We’re going to add. We’re going to click next. All that it is asking me is for the host name of the incidence on the other side. I’ll go ahead and act properly there, and then it’s going to prompt me for the password for the other side.
We go ahead. Just with that click, I’ve made my system more resilient. I’m going through the process right now of replicating my data into a different availability zone.
Given my being able to do a manual failover, if push came to shove, to be able to access my data from a different node. If you notice, we’ll go ahead and we’ll show that during this process, we are now taking everything that we created, even the readable/writable close, and we are moving that over to the secondary instance.
I’ll show you. We have source node, we have a target node, and it’s being asynchronously replicated between nodes. This is great. We have the ability to have our data in two locations. It took little or no time in order to do that.
What if we could stand up two instances but put a VIP in front of those instances and give you the ability to failover in case of an issue. We have the ability to do that with snap HA.
We are going to select our virtual IP. You have the ability to use a virtual IP or an elastic IP. We are going to select a virtual IP. For the sake of this demo, we are going to ahead.
I’m going to choose 220.127.116.11. I’m going to select next. Because of the IAM Role we don’t need to share out the access key. We are going to go ahead and click install. While we are going and installing this, there is a bunch of heavy lifting that’s being done in the background that SoftNAS is handling on its own.
We are updating routing tables. We are ensuring that S3 is accessible. We are making sure that each instance could actually talk to each other. We are setting the heartbeats and putting the heartbeats in place.
With that VIP that sits in front of these instances, we want to make sure that if your source node goes down that your target node is going to stand up and be able to give you no downtime access to your data.
We are going to have around 30% for a couple of seconds. We are going to jump to 50, and within a second, we’re going to be at a hundred. There we go. This is real-time. It was not sped up, just to show how easy it is for us to be able to set up a HA environment which allows you to be able to, from a cloud standpoint, lift and shift your applications. Giving them the ability to use the protocols that they are familiar with. That’s your NFS, that’s the CIFS, AFP, and iSCSI if you see fit.
With that, I’ll end the video showing part of the demo and I’ll go back to showing more of the PowerPoint slides. I’m sorry. Let me take a sip of water. I apologize.
I know if you’re in a room with a bunch of folks or even if you’re by yourself, you’re saying, “It looks good but how much is this going to cost?”
I am actually very glad that you ask that question. I want to introduce you to the TCO Calculator. You could use this calculator to compare the costs of your running applications with what it would actually cost you within an AWS environment.
All you have to do is describe a bit of your configuration and you’d be able to do it. Let me go ahead. I’m going to show it fairly quickly. Let’s calculate. It’s very easy.
You’re going in. You’re selecting currency. It’s going to ask you what type of environment it is and the region, your application, what’s the amount of VMs that you’re looking at, and storage. What are you looking from a storage standpoint?
For the sake of this demo, we went ahead and we did a three-year total cost of ownership for AWS for 40TB instance. What you could actually see is that there is a huge cost-savings between going from an on-premise system to an AWS system and the majority of this cost-saving is actually located in infrastructure.
Getting rid of the need for you to manage the underlying infrastructure ends up saving you about 61% worth of your costs-savings.
How do you pay for this? Right after how much does it cost? The next thing is how can we pay for it? The challenge is you want to move to AWS but where’s your budget for AWS?
We talked earlier in this slide where budgets are not increasing. Budgets are being struggled, but you’re being forced to think about the factor of moving to the cloud and how to move to the cloud.
What we suggest is an innovative method is to allocate your existing maintenance in you budget to be able to make your shift to the cloud. We do have a mock up right now just in the slide.
We are seeing that for a 50TB on-premise system, from a maintenance renewal budget, we’re looking at around $450,000. For $265,000, you could increase the size of your environment in AWS by 20TB. That in itself is cost effective enough.
At the point you select it. Some of the steps that you might be thinking about. Pick a tier 2 application to migrate to AWS, test the waters with that application. From your learnings, you could then create a workplan to migrate the remaining apps, workloads, and data.
When all the migration is done, it’s time for you to unplug your on-premise hardware and have a party because of all the money that you are going to be saving.
Lift and Shift. SoftNAS allows you to lift and shift while maintaining the same enterprise-level of service and experience that you are used to. We are the only company that gives a no downtime guarantee SLA.
We will give you the same enterprise feel in the cloud that you are used to on-premise. Whether or not that’s serving out your data via NFS for your apps that need NFS, CIFS, or SMB, we have the ability to do that.
We deploy within minutes. We give you the ability, as we demonstrated, to do storage snapshots. The GUI in of itself is easy to learn, easy to use, no training. You don’t need to send your teams back to get training to be able to use SoftNAS as a software.
It is there, it is intuitive, and it’s easy to use. We talked about disaster recovery in HA and being able to move to a monthly or annual subscription.
Help us help you migrate your existing applications to the cloud. We allow you to use the standard protocols. We leverage AWS’s storage elasticity. SoftNAS enables existing application to migrate unchanged.
Our software provides all the Enterprise NAS capabilities — whether it’s CIFS, NFS, iSCSI — and it allows you to make the move to the cloud and preserve your budget for the innovation and adaptations that translate to improved business outcomes.
SoftNAS can also run on-premise via a virtual machine and create a virtual NAS from a storage system and connect to AWS for cloud-hosted storage.
Taran: Thanks for that, Kevin. Going on to our next poll. I’m going to launch that real quick. What storage workloads do you intend to move to the AWS Cloud? For those of you who are interested in moving to AWS, is it going to be NFS, CIFS, iSCSI, AFP, or you are not just intending to move to AWS at all?
It looks like we’ve gotten about half of you to vote so far, so I’ll give that about 10 more seconds. I’ll go ahead and close the poll and pull up our results. It looks like about over 40% of you are intending to move NFS workloads to the AWS Cloud, which is pretty common with what we’ve seen.
CIFS and iSCSI support is also balanced too. A couple of you just have no interest in moving to the AWS Cloud. For those of you who don’t have interest in moving to the AWS Cloud, again, in the questions pane, please let us know why you don’t intend tomove to AWS. I’ll go ahead and hand it over back to you, Kevin.
Kevin: Not a problem. Let’s go over. I’ll do a very quick review of SoftNAS. SoftNAS in a nutshell is an enterprise filer that exists on a Linux appliance with a ZFS backing in the cloud or on-premise.
We have a robust API, CLI, and cloud base that integrate with AWS S3, EBS on-Premise storage, and VMware. This allows us to provide data services like block replication, allows you to access cloud disks.
We give you storage enhancements such as compression and in-line deduplication, multi-level caching, and the ability to produce writable snap clones, and encrypt your data at rest or in-flight.
We continue to deliver best of brand services by working with our industry-leading partners. Some of these people you guys might know such us Amazon, Microsoft, VMware. We continue to partner with them to enhance both our offerings and theirs.
These are some of the brands that you trust and they trust us also because these companies are using SoftNAS in many different use-cases all over their environment.
We have a couple of these just to bring out a few – Netflix, Weather Channel, and too many to name.
Taran: Thanks, Kevin. I’m just going to go ahead and take over the screen share really quick. To thank everyone for attending this webinar, we are handing out AWS credits. If you click on this Bitly link right here, you’ll actually be able to register for a $100 AWS credit.
We’re only giving out 100 of these so the first 100 people to register for it will get it. I’m also going to paste the link here in the chat window and you’ll be able to click on that link to register for the credit.
You just have to go to the page, put in your email address, and one of our team members will get back to you later today with your free credit information.
As far as next steps go, for CXOs in our audience, we invite you to try out that AWS TCO Calculator. If you have any questions about it, feel free to contact us. Just visit our “contact us” page here on the bottom and we’ll be happy to answer any questions that you might have about cloud costs and how SoftNAS and AWS result in more cost savings over an on-premise solution.
We also recommend that you have your technical leads try out SoftNAS. Tell them to visit the softnas.com/tryaws page and they’ll be able to go and try out SoftNAS free for 30 days.
For some of the more technical people on our audience, we invite you to go to our AWS page where you can learn a little bit more about the details of how SoftNAS works on AWS. Any questions that you might have, again, please feel free to reach out to us.
We also do have a whitepaper available that covers how the SoftNAS architecture work on AWS and that gets into some technical details that should help you out.
That covers everything that we had to talk about today, so now let’s go on to the Q&A session. We do thank you guys for asking questions throughout the webinar. I’ll go ahead and start having a few questions answered.
Our first question here is why chose SoftNAS over the AWS storage points?
Kevin: I think that that’s a good question. Why? We give you the ability, from a SoftNAS standpoint, to be able to encrypt the data. As we spoke about in the webinar, we’re the only appliance that gives a no down-time SLA.
We stand by that SLA because we have designed our software to be able to address and take care of your storage needs.
Taran: Thanks for that Kevin. Out to the next question — is Amazon the only S3 target or are there other providers as well?
Kevin: Amazon is not the only S3 provider. We do have the ability to connect to blob storage, and we are in other platforms other than AWS – that would be CenturyLink, Azure, among some of the others.
Taran: Thanks Kevin. Our next question; do you have any performance data for running on standard EBS non-provisioned volumes?
Kevin: I think that there might be a session. I believe that question came from James. James, feel free to reach out to us. We’ll definitely get on a call with you, and that’s something that we could dig into more as we discuss your use-case. Depending on your use-case and depending on the enhancements that we could actually put in front of these disks, we could definitely come up with a solution that would be beneficial for you.
Taran: Thanks Kevin. Then onto the next question. If I move my storage to the cloud, do I have to move SSL too?
Kevin: It totally depends on your use-case. This is another use-case question, and depending on your internet connectivity, that’s something that would be defined by that. The last mile of your connectivity into the cloud is always what’s going to determine your level of performance.
Each use case is different, and that’s something that we could definitely do a deep-dive and dig into your use-case again.
Taran: Thanks Kevin. It looks like we have two more questions. We’ve got quite a bit of time so for those of you who are interested in asking questions, just go ahead and paste them in the questions pane and we’ll be happy to answer them.
As far as our next questions goes, what kind of use-cases are best for SoftNAS on AWS?
Kevin: That’s a very good question. There are multiple use-cases. If you have production data that you need CIFS access for, SoftNAS gives you the ability to do that. You have a web back-end that you need to be able to apply iSCSI or more performing disk. We give you the ability to do that also.
Like we talked about in this webinar, you have an application that you need to be able to move to the cloud. However, in order for you to rewrite that application, it’s going to take you six months to a year to be able to support S3 or any kind of block storage.
We give you the ability to migrate that data by setting up a NFS or a CIFS [inaudible 50:29] whatever that application is used to already in your enterprise.
Taran: Thanks Kevin. It looks like that’s all the questions that you guys have today. Before we close up this webinar, we just want to enlighten you to try out SoftNAS free for 30 days on AWS — just visit softnas.com/tryaws.
Also, at the end of this webinar, we do have a quick survey available. We ask that you please fill out the information so that we can better serve in the future. Just knowing what kind of webinars really interest you will help us out with improving our webinars.
With that, that’s all we had for you today. We hope you guys have a great day.