Show HN: gpudeploy.com – "Airbnb" for GPUs

gpudeploy.com

267 points by nicowaltz 15 days ago

Hi HN,

YC w24 company here. We just pivoted from drone delivery to build gpudeploy.com, a website that routes on-demand traffic for GPU instances to idle compute resources.

The experience is similar to lambda labs, which we’ve really enjoyed for training our robotics models, but their GPUs are never available for on-demand. We are also trying to make it more no-nonsense (no hidden fees, no H100 behind “contact sales”, etc.).

The tech to make this work is actually kind of nifty, we may do an in-depth HN post on that soon.

Right now, we have H100s, a few RTX 4090s and a GTX 1080 Ti online. Feel free to try it out!

Also, if you’ve got compute sitting around (a GPU cluster, a crypto mining operation or just a GPU) or if you’re an AI company with idle compute (hopefully not in a Stability AI way) and want to see some ROI, it’s very simple and flexible to hook it up to our site and you’ll maybe get a few researchers using your compute.

Nice rest of the week!

ezoe 14 days ago

If you let third party stranger to offer GPU resource, how do you deal with:

1. Privacy. An attacker can set up a GPU honey pot and and sell the data they got.

2. Fake GPU computation. An attacker can fake GPU and send back dummy data sometimes to reduce the computation.

3. Corrupt GPU. Practically same with 2. But not malicious intent. It's just the faulty GPU.

  • p4bl0 14 days ago

    Regarding 2 and 3: depending on the type of computations you need to do there are ways to verify the integrity of a delegated computation with very high confidence for very low cost. Typically if you're computations happens in a modular structure it's quite easy. See for example my THC (trustable homomorphic computation) project and the accompanying paper: https://pablo.rauzy.name/software.html#thc There is also a video presentation because the conference where the paper was published took place during a COVID induced lockdown: https://www.youtube.com/watch?v=6DByVlqpH0s

    • p4bl0 14 days ago

      Can someone explain why a relevant peer-reviewed paper reference got downvoted?

      • zorgmonkey 14 days ago

        Technically something like this is relevant, but most people would assume that in practice it is going to be way too much work to get it implemented on GPUs and in such a way that it doesn't add too much overhead.

  • evilduck 14 days ago

    Two and three are really the same problem and are solved by periodically running the same workload in duplicate on other systems and comparing results to detect unreliable sellers. Same as all the captcha systems. It's an overhead required to have trust built into a system. In the event of a faulty GPU, the owner would likely want to know anyways.

    The first issue could be addressed by just never giving a seller enough data from a buyer to abuse, but that does require a critical mass of both buyers and sellers before you could distribute and dilute a single customer's workload enough. The company could bootstrap this with renting their own GPUs and subsidizing sellers at first though.

  • TazeTSchnitzel 14 days ago

    There's a privacy issue in both directions too: the tenant will be afraid their workload will be stolen, and the host will be afraid other data in their system will be stolen (if the sandboxing is imperfect).

  • larodi 14 days ago

    Wait a year and see how kids get on blockchain to sell and buy GPU resources for rendering ‘trans furries’, or better analyse classmates’ stolen chats.

    I can vividly remember we were by no means disclosing private info back in ‘94 only to see a world of influencers in 2024 sharing their very personal guts for profits.

    • Foobar8568 14 days ago

      It's only an extension of big brother from the late 90s, to not say real world from MTV in the early 90s.

      At the end of the days, everyone wants and hopes to be an idol or famous.

      • larodi 12 days ago

        ...but also at no point in known history has it ever happened to find millions of people doing so in very mean and aggressive way, and worldwide scene to exploit.

        agreed it IS a Big Brother of a sorts, where everyone participates to certain degree. but then written forums have been a thing before BigBrother and perhaps don't have this voyeurism added in the equation.

  • mnahkies 14 days ago

    I know Azure has a confidential computing offering for GPUs, which I'm hoping will get broader uptake soon. It seems like the best way to address these concerns to me.

    https://azure.microsoft.com/en-us/blog/azure-confidential-co...

    • aliljet 14 days ago

      how would you propose this might work?

      • mnahkies 14 days ago

        I'm not an expert on the area, but I've attended some conference talks on the subject at fosdem so I'll give it a go.

        Essentially you're trying to provide a way to prove that the code running on the machine is what you instructed. This is achieved by a series of hardware attestations that measure and check the code to make sure it's what you requested. Generally this means encrypted ram at a minimum, and checks/balances that give you confidence this is the case (you have to trust someone, eg: Intel)

        Is it perfect, probably not, but it's a lot better than just running VMs with unencrypted memory that any operator can jump into.

        To my understanding most GPU workloads are not run in this way currently, and the operator can see/manipulate everything executed

yeldarb 15 days ago

Love the concept.

I've used vast.ai (similar "Airbnb for GPUs" pitch) for years to spin up cheap test machines with GPUs you can't really find in the cloud (and especially consumer-grade GPUs like 4090s). Any insight into how this is different/better?

  • nicowaltz 14 days ago

    Main difference is that we are more opinionated (in terms of configurations) and sort of do the scrolling and sorting out for you – hopefully a bit smoother as a user experience. We sort out bad machines immediately. We're also directly working on making compute from unknown high-end data centers available, there's a lot of unused compute out there! See gpulist.ai

    Also, don't know if vast.ai does this, but with us you can have 6 user sessions on your machine if you have six GPUs, so granular utilization is possible.

    • bradfox2 14 days ago

      We own and operate 40+ data center GPUs (v100s, a100s, and ax000s) in a private cluster and use vast to rent unused capacity.

      What would make you better than vast is extremely easy spot leasing and job prioritization.

      I want to be able to have one of our training jobs finish, and then have the capacity immediately transition to a lease. With vast, we are renting in week long blocks.

      • jcannell 14 days ago

        You should be able to do that right now on vast. You just need to rent the gpus yourself with your own on demand instance(s) for your training job. As soon as it finished you then stop or destroy those instance(s) and the GPUs are available immediately (and if there are any other instances queued up in scheduling they will start up). Your actual job doesn't necessarily need to run in the container (if you know what you are doing).

        (I'm the founder of vast btw - contact us for help on setting this up and/or any feedback on making it an easier/better process)

      • nicowaltz 14 days ago

        Exactly, that's the idea

        • bradfox2 14 days ago

          Is it implemented?

      • DeathArrow 14 days ago

        Can you expand on spot leasing and job prioritization? What kind of api would you prefer? How would you like to adjust time slices?

      • bigcat12345678 14 days ago

        You can do that on llm.sxwl.ai Shoot me an email at z@sxwl.ai for instructions, the web site is pretty outdated, the main UI is through restful API (which we don't have time to write doc yet)

        • mkl 14 days ago

          If you have time to answer emails with instructions, you have time to update your site and documentation. Why not just do that?

          • grepfru_it 14 days ago

            Don’t build something if you don’t have a use case. All you have is a wishlist, until someone says “yes I want this here is $$$” which I assume the email will facilitate.

    • cfn 14 days ago

      Just a quick comment: The country list in your Add Basic Node Data is not sorted.

  • ganoushoreilly 15 days ago

    Im also interested in what the differing factor is. Would also like to see more documentation for onboarding rather than just "Ubuntu and root available".

    • icelancer 14 days ago

      vast you have to choose specific machines. gpudeploy routes to whatever resources are available.

      vast has a lot of bad machines with terrible PCIe lanes and architecture you have to learn the hard way. Someone on HN wrote a script to run a test docker image on every machine and auto-tagged the machines' quality using their API, which is what I'd do if I was going to use vast seriously for compute.

      • lelanthran 14 days ago

        >> Im also interested in what the differing factor is.

        > vast has a lot of bad machines with terrible PCIe lanes and architecture you have to learn the hard way.

        Wouldn't gpudeploy have exactly the same problem? How is it mitigated with gpudeploy?

        • everforward 14 days ago

          I think it's more of a business strategy issue than a technical one.

          I suspect it would be trivial for Vast or GPUDeploy to spin up a benchmarking job before allowing sales on that machine. I'm not an expert on PCIe lanes, but I would think the performance issues would be visible via bandwidth or latency on the lanes.

          It kind of makes sense to me, though. If I were looking for absolute reliability and was willing to pay for it, I'd just go to one of the many GPU cloud vendors. Likewise, I suspect anyone willing to really work on getting good performance would rather be a real provider or sub-provider than being part of this nebulous C2C GPU cloud.

      • polygot 6 days ago

        Do you have a link to that thread/Docker image? I would be very interested using it

  • firloop 15 days ago

    I use vast.ai somewhat often. It's great!

  • tehsauce 14 days ago

    +1 for vast. they usually are the cheapest and have the most supply. some instances can be less reliable at the low end though

nirvanatikku 15 days ago

Great idea and wishing you the best of luck!

Dropping a note that I've found https://akash.network/ ~ https://akash.network/gpus/ to be impressive, as typically with crypto projects it's all scams, however in this instance there's demand and legit usage. https://stats.akash.network/

Something to consider!

  • lordofgibbons 14 days ago

    I was excited to see their offering, but unfortunately they don't take fiat - only some specific cryptos.

    • barbarr 14 days ago

      I feel they could onboard so many more users by having a fiat frontend that doesn't refer to blockchain. But maybe they're targeting crypto natives so users don't get shocked by price variability etc.

      • david_shi 12 days ago

        you can use akash gpu's on brev.dev

spencera 14 days ago

I like how signing up requires agreeing to the terms of service and privacy policy, but both links go to empty pages.

  • nicowaltz 13 days ago

    These links should work now, at least they do for me. We use a third party service called Termly for these, I think there was an issue with the url (missing the www).

ericyd 14 days ago

Might just be me but I think "On demand GPU provisioning" is easier to understand than "Airbnb for GPUs"

  • montebicyclelo 14 days ago

    From the YC guidelines:

    > One good trick for describing a project concisely is to explain it as a variant of something the audience already knows. It’s like Wikipedia, but within an organization. It’s like an answering service, but for email. It’s eBay for jobs. This form of description is wonderfully efficient.

    Seems the risk of this is a loose simile.

    Edit: Though, thinking about it, Airbnb for GPUs is fairly accurate for this model of a marketplace where people let out their GPUs for others to rent.

    • michaelt 14 days ago

      > Airbnb for GPUs is fairly accurate for this model of a marketplace where people let out their GPUs for others to rent.

      Minimum rent 1 day? $100 cleaning fee? No quality control? Spy cameras in the bedrooms?

    • ericyd 14 days ago

      Yeah I think in general it's a decent strategy. I'm this case it falls short for me because a key element of Airbnb is that the assets (homes/rooms) are provided by the user base, whereas this product doesn't seem to do that at all. So the strategy of using familiar companies is good, but I think it only works when the functional correlation is strong. In this case I don't think it is.

    • Volt 14 days ago

      The audience here already knows what "on-demand provisioning" means.

  • wg0 14 days ago

    On demand means you're another underdog that's just like Google or AWS offerings which might not be around next month.

    Airbnb implies a two way market place. Much more interesting prospect for those having idle GPUs .

    Not sure if electricity costs justify renting out and what pricing looks like, those details apart.

    • bobbob1921 14 days ago

      More so than the electricity of the GPU, is the electricity in cooling the GPUs heat. (In most cases cooling, in some cases, I’m guessing the heat might be a welcome bonus but that’s more rare)

  • nojvek 14 days ago

    The issue with AirBnb is that it doesn’t have the best reputation nowadays.

  • kobalsky 14 days ago

    on demand gpu provisioning misses the part that you can sell your gpu time, doesn’t it?

latchkey 14 days ago

Interesting, these companies are springing up left and right. Software solutions to a hardware problem. It looks like most of the available compute is allocated, so they are going to have to get a bunch more providers onboarded.

I wonder why y-combinator is stuffing their investments with multiples of these similar companies... https://www.shadeform.ai/ is another one.

A few quick comments:

Reading the source of their install script:

https://gpudeploy-public.s3.us-west-1.amazonaws.com/join_clu...

It doesn't start off with set -e, which could result in an incomplete install, yet appear to finish.

It also installs some binary "instance-server"... who knows what it does... would you trust this on your server on your network?

It is nvidia specific... sadly, don't expect AMD gpus anytime soon.

Feels like a MVP, let's see how this grows over time.

  • dheera 14 days ago

    Is there any way to monitor for sales of hardware if/when any of these companies dies so that I can get 1xH100 for personal use at a deep discount?

    • latchkey 14 days ago

      Ebay?

      • dheera 14 days ago

        Is that where they really go? I feel like buying anything for several thousand dollars on Ebay is sketch. What if they ship me an empty box?

  • tptacek 14 days ago

    As the post points out, this is a pivot from a company that was a consumer of this kind of service in its original incarnation; YC had presumably nothing to do with it.

    • latchkey 14 days ago

      That an interesting presumption. If a startup is "Backed by Y Combinator", and it pivots, you don't think that Y Combinator would have some sort of opinion on that?

      I would expect some sort of conversation like: "Oh hey, we have another company, in our portfolio, that is far ahead of you, doing exactly the same thing. Maybe you should do something else?"

      But of course, this is AI... plenty of space for gpu marketplaces.

      • dang 14 days ago

        YC partners might have an opinion and might offer advice but they're always going to support what the founders want to do—that's kind of the core principle of YC.

        As for "other company in portfolio", that's unavoidable when funding thousands of startups and almost always turns out to be a non-issue.

        • latchkey 14 days ago

          Awesome, makes perfect sense. Thanks for the clarity @dang.

          I'm curious now, if you can say. Was advice offered in this case? If so, what was it?

          • dang 14 days ago

            I'm not part of such discussions so I'm afraid I have no idea!

            • 2024throwaway 13 days ago

              People here think you _are_ YC. Little do they know you're far more perceptive and capable than YC.

              • dang 13 days ago

                That's definitely not true, but thanks for the support :)

                • 2024throwaway 13 days ago

                  Truth is a spectrum. I will leave my opinions on YC unpublished. Thanks for all you do.

      • tptacek 14 days ago

        That's not how YC works. We're YC W20. We have a GPU offering. Guess how many conversations we had with YC about it.

        What's YC going to do? They have a tiny stake in your firm. That's the point; that's what "founder-friendliness" means. They're not your board.

        • latchkey 13 days ago

          Cool, thanks for the info Thomas. I honestly have never looked into YC at all, so I'm definitely not aware of how these things work. I was posting from an honest curiosity.

  • moneywoes 14 days ago

    why does ycombinator invest in companies that will compete against each other, isn’t that a conflict of interest

    • gadflyinyoureye 14 days ago

      Nope. That allows them to control the market. Find new ideas that could threaten the big 5 status quo. Fund them. Get them bought by one of the big five (making a good amount of money off that). Let the big five shutdown the competition.

      Conflict of interest only applies if you think they’re concerned with the public good. There is no assumption of that here. Every man for himself.

bllchmbrs 15 days ago

For those that wants prices checkout: https://gpumonger.com/

This seems much more in depth, and a true service, but for those that just want to compare prices check out gpumonger.

  • latchkey 14 days ago

    There is no 'contact us' page, which is a little concerning.

    • Ocha 14 days ago

      They are not providing actual service - just price comparisons across different providers. Why is contact us page missing a concern here?

      • latchkey 14 days ago

        I'm a provider, I'd like to list my service. How do I do that? Since there is nothing obvious, then how do you know the providers they list are a valid representation of what is out there? Maybe there are other providers who don't want to pay to be listed? But how do you even pay?

        • trogdor 14 days ago

          I have read many of your comments on GPU-related posts. Occasionally, when I see your comments, I visit your website to see if I can learn anything about your business beyond what you are posting on HN. To this day, hotaisle.xyz contains nothing but a logo.

          I don’t get it.

          How can you expect a website to list your service when your own service’s website contains zero information? Why would you pay to list your service, when there is no information available about the service you provide? Am I looking in the wrong place? So confused…

          • latchkey 14 days ago

            > hotaisle.xyz contains nothing but a logo.

            To be fair, there is also a contact email at the bottom of the page.

            > How can you expect a website to list your service when your own service’s website contains zero information?

            We are the first and only (for now) verified MI300x provider on gpulist. In order to get verified, I contacted them directly, they asked for a few bits of information about my business, including my EIN. What I'm offering there, is exactly what I have today.

            https://gpulist.ai/detail/3c18f8a

            > So confused…

            I know, it is ok. Let me explain a bit. We are starting small, so the website is the last focus right now. I know the general expected culture is to have some splashy page with a typeform on it, but hey... aren't we also a bit tired of that?

            In order to even get access to buy these GPUs, you have to go through quite a lot of effort. You can't just buy them off the shelf from BestBuy. They are export controlled and I've agreed to not use them to build bombs. You have to have a valid business and a great story, or they won't even talk to you. Heck, I even had to prove my business was in good standing in Delaware. I'm pointing this out because I will need to know all my customers too. My business isn't something someone just signs up for on a website.

            These GPUs are also extremely expensive. Imagine a 350lbs Ferrari. We started with 8 of them (one chassis) because they are super new and it was a proof of concept. Last year, we didn't even know for certain if AMD would double down on AI. This is all we initially raised funding for. As soon as we deployed the compute, we immediately had a customer on them, all without a website. Just word of mouth. By the way, the success of the PoC unlocked our next round of funding, and we are working on a much larger order of MI300x right now.

            Don't worry, you'll get a website at some point. That said, these things sell themselves, you either have them or you don't. I've been very transparent and public about what we are up to. Would a website really help here? Maybe. But I've also started other extremely successful businesses originally without websites too. At the end of the day, I'd rather spend investors money on buying more compute, than a pretty website. Once I have some more revenue, I'll funnel that right back into the business and work on marketing/sales more.

            If you're curious about anything, feel free to just reach out and ask. I'm not some corporate overlord suit wearing sales guy. I'm an open source tech nerd who's been in the business a long time. 20+ year ASF member, who co-founded Java @ Apache. Happy to answer any questions.

            • trogdor 13 days ago

              Good luck with your business!

              • latchkey 13 days ago

                Thank you, we need all we can get! Super excited about it though. Big things coming in the pipeline.

        • mike_d 14 days ago

          They list over 40 vendors. If you want to be listed, maybe get bigger?

          • LoganDark 14 days ago

            Who says they're even looking? Being able to contact them would tell you right away whether they simply overlooked you or whether you don't yet meet some criteria (which they might actually tell you).

    • bllchmbrs 14 days ago

      This was my project that I recently sold. If you want to be listed I can put you in contact with the folks that bought it. LMK!

      The lack of "contact us" was pure laziness at the time.

      • latchkey 14 days ago

        Congrats on the sale! Interesting to make a few pages and then sell it.

        Sure, assuming the site will continue... my email is in my profile.

keefle 14 days ago

Neat Idea!

I was wondering, are there any security guarantees for the providers? Assume I have a small GPU cluster at home, if I rent out my GPUs, what sort of access should I accept the renters to have? Only GPU kernels would be sent to machine? Or will they have a limited permission user access on my cluster?

Also instead of having the operatos openning ports in their routers, were there any considers of adding them to a private network for a more seamless experience? (Nebula/headscale and the likes)

malfist 15 days ago

I'm curious how a business pivots from drone delivery to GPU Airbnb?

  • exe34 15 days ago

    I got it from the line about how they trained their robot models - think of Amazon pivoting to aws, the substrate becomes the commodity. Unless you're asking if it makes sense, or what went wrong? Then I don't know. I imagine startups that survive usually pivot from something else?

    • MikeTheGreat 15 days ago

      I've never run a start-up (successful or otherwise :) ) but I've also heard that being able to pivot can be really useful - Flickr is often cited as an example.

      Also, "GPU on demand" sounds _a_lot_ easier than "drone-based delivery". Between Seti@Home/Folding@Home/etc, various grid-computing/clustering/orchestration stacks that already exist, etc it seems reasonably doable to implement in a year or so. "drone-based delivery" sounds capital-intensive, sounds like you'll need to spend a lot of time building a professional network of business people who might use the service (so there's a 'cultural friction' between techie founders and business folks, potentially), plus the ever-looming threat of Amazon/etc figuring this out first.

      tl;dr: I agree it's weird pivot, and good on the founders for being able to make the change! :)

      • nicowaltz 14 days ago

        While most people think it's cool and some think it's scary, drone delivery is ultimately something that people (that is: companies with money) don't need.

        Turns out, using drones to deliver is also not that competitive either. Delivery vans are very cost-efficient, and for food delivery / on-demand delivery, drones are not able to carry most orders. So it's not even the regulatory pains that make this difficult, which are unbearable in their own right.

        It was a lot of fun to work on and we would have definitely stuck with it if there was any interest. There was none, so we had to admit that to ourselves.

        This is a hard pivot, but it's been very stimulating to work on.

        • qeternity 14 days ago

          Btw thanks for this honest insight. When I read the bit about your pivot, I sort of rolled my eyes. But this is a really nice, sober reflection that actually builds credibility with people like me who might make snap judgements at first.

          I hope you guys find traction.

      • jayyhu 14 days ago

        To add to "capital-intensive": the regulatory framework right now (in the US) for drone delivery is uncertain at best. While most of the rules & regs to enable drone delivery is already finalized & published, a few very important ones (such as BVLOS - beyond visual line-of-sight) have yet been finalized, and thus need one-off waivers from the FAA to allow. And getting those one-off waivers is what eats a lot of time & money.

  • callalex 13 days ago

    Their parents are good friends of someone else’s parents.

  • kristopolous 15 days ago

    They're chasing money. It's a tactic

EE84M3i 15 days ago

What's the security story here?

  • remram 14 days ago

    What's preventing GPU providers from sending wrong results instead of actually running the computation? For example, send the last computed result? Is this something that the renter has to handle by adding their own checks?

    In addition to the problem of the renter crashing your machine or reading your password through DMA, of course.

    • latchkey 14 days ago

      What incentive would a GPU provider have to spend time figuring out what result to send for some custom application?

      • remram 14 days ago

        The incentive is huge, if I spend 2 milliseconds sending you your previous results instead of 2 hours running your new computation, I can (pretend to) run way more computations on the same hardware and collect hundreds of time more money.

        • latchkey 14 days ago

          At the risk of being exiled off the platform and earning nothing. Don't forget, there is a bit of KYC with Stripe.

        • krapht 14 days ago

          ID verification before you can host and random audits from gpudeploy.

          • greenish_shores 14 days ago

            NO. That's the worst way to do almost anything on the Internet, and should be considered a last-line defense, if nothing else can be done. Here, it can be. See my comment above.

          • remram 14 days ago

            That's my whole question, do they do random audits, or is it the job of customers to double-check their results for possible attack or compute-theft and report.

            • greenish_shores 14 days ago

              It seems wrong to call it a "job of customers". It's like you wrote a Bitcoin client which didn't verified hashes of transactions, "trusting" everything. Or like serving a website with login feature supporting only HTTP, not HTTPS. It is a very basic feature of whatever software would connect to such services.

              • remram 14 days ago

                So it is the job of the customer to write their own Bitcoin or HTTPS client, in your metaphor.

                • greenish_shores 14 days ago

                  Every technology was (very) underdeveloped at some early point in its evolution.

                  • remram 13 days ago

                    That's what I asked, how developed is it now. Why so defensive?

                    • greenish_shores 13 days ago

                      I don't know how developed is it now, I'm not associated with the startup shown in any way. It's mainly a question to them. However, in terms of wider industry, in general distributed high-performance GPU(-like) computing "for everyone" is in its infancy. 99% of what was already done up to this point was targeted to people who would both buy and supply power "in bulk", not "in retail". Perhaps with a little exception of several excellent projects like Folding@Home and other @home's.

      • greenish_shores 14 days ago

        Run 1/10,000 - 1/100,000 of computations locally, and also send them as tasks to be send remotely. If compare yields difference, repeat both. After, say, 10 tries, blacklist the provider. Of course it will take a lot more nuances to implement that, but that's the general idea. It's a no-brainer.

        • remram 14 days ago

          Sounds like a lot of work that I would expect the paid service to help with.

          • greenish_shores 6 days ago

            Yeah, it's "they" who should do that, of course.

    • greenish_shores 14 days ago

      Linux supports IOMMU on most platforms.

      • remram 14 days ago

        I fail to see how this relates. If you can't trust the provider, why does it matter whether they say they have IOMMU or not.

        • greenish_shores 14 days ago

          It relates to the "get your passwords over DMA part".

  • lschneider 14 days ago

    At the moment, we manually verify operators and are currently onboarding some tier-4 operators. Down the line, we'll have a 2-tier system where you can choose whether you want a verified machine or not. From the operator's perspective, everything runs inside Docker, configured with security best-practices.

    • lolinder 14 days ago

      I've always understood that containers are not proper sandboxes and shouldn't be used for containing untrusted code, no matter the best practices used. Has this changed in recent years? Do you have documentation for what sorts of best practices you're using and why they are sufficient for executing untrusted code?

      • gavindean90 14 days ago

        You are correct from my knowledge. I would expect that if the container is set to not run as root you might be able to enforce fine meaningful security but I’d still run it in a VM if feasible.

    • janosdebugs 14 days ago

      Having done a little bit of work in the area[1], I think you should publicly document exactly what those best-practices are. Are the workloads running in a networkless container? Do you limit IO? Do you limit disk usage? Answering these in detail would help you gain customer trust on both sides.

      [1]: https://containerssh.io/v0.5/reference/docker/#securing-dock...

    • flaminHotSpeedo 14 days ago

      So you don't have real security for operators, is what you're saying.

      Containers are not, and will never be, a secure isolation boundary.

  • htrp 15 days ago

    probably very basic... so don't run it on anything that has your own data on it (if you're an AI startup, definitely don't run it on your research cluster).

    • Traubenfuchs 14 days ago

      > definitely don't run it on your research cluster

      ...what‘s the threat, actually? GPU time sellers stealing your secret sauce?

      • lolinder 14 days ago

        I think they mean don't lease out your research team's GPUs and allow random people to run untrusted code on your cluster, lest they figure out a way to break out of any sandboxing the software has in place and get loose in your network. The company's current answer to that concern is "everything runs inside Docker, configured with security best-practices", which is less than inspiring.

        https://news.ycombinator.com/item?id=40261591

cdchn 14 days ago

If I got an RTX 4090 sitting around idle, how much can I get for it?

  • ant6n 14 days ago

    That’s what I wondered, too. I feel the pricing page should be like a currency trader, showing a list of sell at/buy at.

  • qeternity 14 days ago

    You should be able to get 40-50 cents/hour.

    • poooogles 14 days ago

      About half of that will go to electricity here in the UK (23.72p per kWh). After costs that's roughly $145 a month profit. Might be more worth it if you have rooftop solar and have free electricity during the day.

      • psini 14 days ago

        Or for "free" heating in the winter!

idiotsecant 15 days ago

This is an interesting idea but it would be cool if it were more granular, like I pay while my payload is executing only, and an API abstracts away which GPUs I'm running on, the execution environment, etc and just let's me push in code and get out data when it's done. Maybe that's what this is, not sure. I hit an account sign up before I could figure it out.

  • dbish 14 days ago

    I think products like this have to evolve into pay by the hour otherwise you find alternatives that have it.

IndigoIncognito 14 days ago

Tried deploying an instance, didn't work, no option to let me delete my billing info/ account. Identical services with better features and more availability exist

  • IndigoIncognito 14 days ago

    PS: It was a temporary server side caching issue

scottfits 14 days ago

I have a cluster on RunPod and it's great, but there's definitely some opportunity in the space, for example if you can focus on transparency (utilization rates, more info about end user and their purpose, reporting) that would be helpful.

I think RunPod and Vast have most the market share but t's still early to the game

wg0 14 days ago

> "if you’re an AI company with idle compute (hopefully not in a Stability AI way)"

What are they referring to? Stability AI over provisioned hardware or something? Sorry not up to date with what's happening here, have developed a blindspot for AI hype.

callalex 15 days ago

Back in my day we cooperated to advance medicine and science with folding@home. How times change.

https://en.m.wikipedia.org/wiki/Folding@home

  • EduardoBautista 14 days ago

    folding@home was used in the “console wars” to justify one’s choice of the PS3 over the Xbox 360.

    Was the cell processor in the PS3 really that efficient for this purpose?

    • TkTech 14 days ago

      It was a fairly unique architecture that had some pros and cons. In reality what made it a winner was the steep subsidy by Sony (units were sold at a loss, making the assumption you would buy a few games) and the availability (later removed) of an official Linux distribution.

      • luigi23 14 days ago

        very few cared about linux. it was mostly hd-dvd (later dvd) vs blu-ray.

    • kuschku 14 days ago

      It was basically an all-purpose vector computing monster focused on SIMD. You could use it for physics simulations, animations, tesselation, etc. Basically everything you'd use a compute shader for nowadays.

      That's why emulators need AVX512 support to match the PS3. It was incredibly powerful.

      Obviously, in that era's single-threaded world no engine could make use of that functionality and few knew how to program for it. It was ahead of its time, by quite a while.

      • fl0ki 11 days ago

        Remember around that time when the industry said that OpenCL would allow write-once run-everywhere compute code for a booming industry of diverse and competitive compute devices? I fell for that scam for a couple of years, before very fortunately getting a different job for long enough to watch the collapse from a safe distance.

  • AbstractH24 14 days ago

    Thought the same thing when I read this.

    How times have and haven't changed.

DeathArrow 14 days ago

What are the benefits over Salad or Vast.ai?

If I pause a GPU instance will I be able to resume it on the same host so I can use the data generated before?

CaptainFever 14 days ago

What about privacy? Can I run sensitive data on these strangers' GPUs?

azinman2 14 days ago

Seems like you have to rent for specific dates… that doesn’t feel very on-demand to me. It also means you have to know how much time you need. I’d love to be able to run a job with a remote gpu as if it were local, spinning up instances as needed. Then if things crash and I need some hours to figure it out, let me downsize it all.

oefrha 14 days ago

I don’t get “final rate is usually lower”. Why is it lower, and by how much? And since it’s “usually”, what about the unusual case? Equal to the sticker price, or higher? But since you use “<” neither should happen? I would expect an exact price tag for a “no-nonsense” pricing scheme. Right now it’s just confusing.

  • nicowaltz 14 days ago

    Thanks for the input, you're right, maybe a fixed price is better. The idea currently is that the quoted price is an upper limit on the rate you pay and the final rate depends on what the GPU provider inputs, which will normally result in a lower rate than the upper bound. At the moment we have upper bounds if you want to be included in on-demand input, and also depending on the service you provide (security, network etc.). We determine pricing with partners 1-on-1 a the moment.

    • latchkey 14 days ago

      I would focus on offering a pricing model that entices people to want to pay.

      I've found that usually this involves something extremely simple and easy to understand, with clear up front numbers.

    • oefrha 14 days ago

      Thanks for the answer, probably a good idea to include some version of this in the footnote.

spxneo 15 days ago

doesn't seem all that cheap compared to whats available

a2128 15 days ago

What's so unique about this compared to just using RunPod?

DeathArrow 14 days ago

I like runpod the most. They offer attached network drives so you can start a GPU instance, stop it, and restart while retaining the data. This way you don't have to sync data to and from external data sources.

If I want to run stable diffusion on other provides I have to install docker image, get data from somewhere, run models, sync data to somewhere so I don't lose it. The process can be tedious and time consuming.

thih9 14 days ago

> or if you’re an AI company with idle compute (hopefully not in a Stability AI way)

What does that mean? I’m likely missing some context, could anyone explain?

  • burkaman 14 days ago

    I think Stability is not doing well and ran out of money to pay for all their compute, so they should probably use some Airbnb guests right now.

mattxxx 15 days ago

Doesn't this just raise the price of gpu's, and further encourage people to hoard them?

cfn 14 days ago

How much do these kind of services pay the GPU owners? Before I sign up and add my nodes I need to know if it will be worth it considering I'm paying the power to run it, depreciation, etc.

renewiltord 14 days ago

Hmm, interesting. At $0.49/hr I get $0.30/hr profit out of my 4090s. It's more like Uber for GPUs. When I have spare capacity on my 4090s maybe I will look into the service.

overgard 15 days ago

Pretty neat, but is there any Windows or MacOS support planned? I wouldn't mind renting out my GPU when it's idle, but I don't really want to go through the process of dual booting etc.

  • nicowaltz 14 days ago

    Windows support is definitely something we are thinking about.

  • bongodongobob 15 days ago

    What are you doing GPUs that requires Windows or Mac support?

    • zephyrthenoble 15 days ago

      Video games

      • otterley 15 days ago

        Have you looked at Nvidia GeForce NOW? It's like $10/mo for a pretty decent streaming gaming rig. I'm very happy with it - I don't have to deal with Windows and can play AAA games on my Macbook Pro at 60Hz (1080p).

        • erhaetherth 15 days ago

          > on my Macbook Pro at 60Hz (1080p).

          I think you just answered yourself. Some of us like to play games at 4K at 80Hz+, with no subscription fees, no internet bandwidth requirements, no added latency, and ability to mod.

          • hhh 15 days ago

            Yeah, but that has nothing to do with the context of the question. Someone is specifically asking about accessing a GPU over the internet for video games.

      • freedomben 14 days ago

        Nowadays it's rare that a Windows-only game that I want to run doesn't run flawlessly on my Linux machine (through Proton/Wine). I wouldn't recommend going outside of Steam though unless you're willing to do some troubleshooting.

        • 1992spacemovie 14 days ago

          Try running COD Warzone on Linux and not get quarantined to cheater-suspect lobbies.

          • freedomben 14 days ago

            Ah yes, apologies that's an important caveat I forgot to include.

            Games that use anti-cheat are a mess on Linux. I don't play any of those games, and if you do then you're likely to run into some trouble with Linux-only.

            • 1992spacemovie 10 days ago

              Yeah anti cheat sucks overall tbh. I loathe giving kernel level access to a random video game’s anti cheat system.

              Short of dedicated hardware (Xbox/ps) I’m not sure what else could be done.

      • bongodongobob 14 days ago

        So what are you looking for, an RDP session with an attached GPU?

floor_ 14 days ago

I remember gamedevs wanting a company like this to test video game drivers and shader code on a slew of different cards and vendors but this seems clearly set up for LLM work.

KeplerBoy 14 days ago

Make sure your users can run profiling tools! Specifically Nsight Compute needs containers with some elevated privileges. That would get you an easy win over vast.ai .

rushingcreek 14 days ago

This is really great! I hope you succeed. At Phind, we’re customers of Voltage Park and SF Compute for something similar. Let me know if you’d like to chat!

lordofgibbons 14 days ago

Hello, I'm currently in the market for an on-demand GPU host so this is perfect timing. What kind of privacy can I expect from this service?

anshumankmr 14 days ago

If I have a 3060 TI on my laptop, could I somehow rent that out? Just curious. Not that I intend to do that, but wanted to know.

CtrlAltDelete51 14 days ago

What kind of utilization percentages do you expect to be able to provide? Do you have any existing usage/ROI data?

  • nicowaltz 14 days ago

    Just launched but will put a panel on the landing page for that, once the data comes in.

animex 14 days ago

I wonder if I could get strangers to heat my home with GPUs in the winter...

modeless 14 days ago

How are you different from vast.ai? What's your payout rate?

CtrlAltDelete51 14 days ago

Side note: you contact form doesn’t work - do you have an email I send some questions over to?

  • nicowaltz 14 days ago

    Sure, nico@gpudeploy.com

densh 15 days ago

Are you planning on letting hobbyists rent their single machine with multiple GPUs?

  • nicowaltz 14 days ago

    Definitely, but we will sort out machines that are unreliable quite quickly.

anyfactor 14 days ago

I am interested in the business logic of paying to hosts. Say someone had netted sub 10 dollars for hosting their GPU on the platform. How do you make that payment? Business have to pay fixed plus variable transaction fees. So the small payments to a large pool of users should be quite involved specially with international payments.

Since the dawn of torrenting (and to some extent the the Silicon Valley TV show) many folks tried to launch a P2P resource pool platform. The challenge is that small amount of money being distributed to a large pool is challenging. Then there is the issue of stability. Do you penalize hoster who fail to provide a stable service?

Hosters must have "trust" in the system and platform can only afford to pay out only when the payable amount to hosters reach a certain threshold. Like you can't withdraw unless you have about $100 charged invoices.

And that is where I see the concept of crypto tokens appear because trust as a short term mechanism can be consoled via crypto tokens.

Engineering an service to take advantage of idle resources is a non-trivial task specially if the service is hooked up to doing something dedicated. For the hosters energy is not cheap, maintenance is a hassle and guranteeing stability is a pain.

I would love to learn how you would address these issues.

epa 15 days ago

DDoSaaS

moneywoes 14 days ago

what is a Stability AI way?

krasin 15 days ago

Related: https://vast.ai is good and cheap. Just don't put any sensitive data on these GPU machines.

I've been a happy user of vast.ai for some time now.