Is there a Public FTP server to test upload and download? [closed] - download

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 4 years ago.
Improve this question
What I want to do is measure broadband speed using c#.
To do this, I use NetworkInterface.BytesReceived and BytesSent to get the current amount of bytes sent and received from my network adapter and get the DateTime.
Then I download/upload a file and get the BytesReceived/BytesSent and DateTime again. The download/upload speed is just the the difference between the BytesReceived/BytesSent divided by the time difference.
Now I am wondering if there is a public FTP server where I can download and upload files.

Tele2 provides ftp://speedtest.tele2.net , you can log in as anonymous and upload anything to test your upload speed. For download testing they provide fixed size files, you can choose which fits best to your test.
You can connect with username of anonymous and any password (e.g. anonymous ).
You can upload files to upload folder. You can't create new folder here.
Your file is deleted immediately after successful upload.
Found here: http://speedtest.tele2.net/

Try ftp://test.rebex.net/
It is read-only used for testing Rebex components to list directory and download.
Allows also to test FTP/SSL and IMAP.
Username is "demo", password is "password"
See https://test.rebex.net/ for more information.

Currently, the link dlptest is working fine.
The files will only be stored for 30 minutes before being deleted.

There's lots of FTP sites you can get into with the 'anonymous' account and download, but a 'public' site that allows anonymous uploads would be utterly swamped with pr0n and warez in short order.
It's easy enough to set up your own FTP server for testing uploads. There's plenty of them for most any desktop OS. There's one built into IIS, for instance.

I have found an FTP server and its working. I was successfully able to upload a file to this FTP server and then see file created by hitting same url. Visit here and read properly before use. Good luck...!
Edit: link is now dead, but the FTP server is still up! Connect with the username "anonymous" and an email address as a password: ftp://ftp.swfwmd.state.fl.us
BUT FIRST read this before using it

Related

How to access Cpanel on 1&1(IONOS) hosting? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about programming within the scope defined in the help center.
Closed last year.
Improve this question
Does anyone know how to access Cpanel on 1&1(IONOS). One of the biggest in the web hosting market but I didn't know it was one of the worst. Knowledge Base is empty. Searching on Google doesn't result in anything.
I am startled by the fact that I found NO INFORMATION online. Absolutely NO information on google. How could it be possible? No one has ever tried to ask this question before?!?!? BTW IONOS is the worst web host. If someone has to search for 20 mins to access even cpanel, then it's the worst web host.
IONOS Screenshot
Note the below is only available if you have the right build assigned to your account.
Server administration with Plesk is conveniently done right in your browser.
Since Plesk provides its own HTTPS service for this, a web server (e.g. Apache)
does not need to be running on the server.
Open your web browser and enter your domain name (or IP address)
followed by :8443 in the address bar.
Log in with the user name Admin and your initial password.
You can log in to IONOS and find your initial password in the Server Access Data section.
I am sure by now you managed to fix your problem. But I figured I must just post how I accessed it for other people. I am assuming that you wanted to remove some of the files on your website (i.e. if a WordPress plugin destroyed your site). The way I accessed my files is as follows: Firstly go to your Ionos dashboard. Then click on the contract you are using (i.e. WP Business). Thereafter go to hosting and click manage, then a page will open and click webspace. Once you click webspace you will be able to access all of your files.
I hope this helps someone out because truly getting information about Ionos seems like thesis research.
I nearly fell for their $6-for-the-first-year Business Plan account. Whups! Good thing I asked some questions first.
The sales person answered the chat after approx 20 mins, and was not customer friendly. When I enquired about included CPanel apps, I was abruptly told that CPanel is not included - you must purchase it separately. She did not tell me (this is what I mean by "abrupt"), but I read elsewhere, that they wrote their own "CPanel-like" administration tool - but from what I can tell it is significantly less capable. And as of this writing, I still do not know if they offer Node.js (which was something that I asked the sales rep but she did not answer).
I would be interested to see a screenshot of the IONOS CPanel replacement. I also would like to know what included apps they offer, such as phpBB, SimpleMachinesForum, Node.js, Python, Ruby - and how Git management/deployment works. As of this writing, and over 30 minutes on chat with them (including wait time), I have no idea.
Here is a fairly recently article from Dec 2021 that sheds more light:
https://websitesetup.org/hosting-reviews/ionos/
You can access your Cpanel using below URL,
http://IP:2082
Note: In IP you have to set your Cpanel IP.
above link redirect to login page, In login page you need to enter your username and password,
From the look of your screenshot it looks as if the package you have purchased is not one that contains a server. Here are two options for packages with servers.
Cloud
Dedicated
Ref
Go to Servers & Cloud -> Infrastructure -> Servers and select the server you want to access.
Then, go to either the Plesk or cPanel section, there you'll find user, password and host for your server.

How to cache every call made for an offline web experience [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I'm going up to the mountains with no internet connection to present something. I'd like to be able to use interactive examples since I'll be presenting on a certain website.
So is there a way I can set up a proxy caching server or something to cache every call made in order to have a fully cached website experience with no internet connection?
I've looked at http://squidman.net/ but I'm not sure how it works or how to use it.
You might want to try something like this. It might be a lot more work than the steps below, but this could be a good starting point.
Create a local proxy server along with memcache or redis
Update the browser proxy settings to use your proxy server details
Make the local server look for the url in the redis server.
If found, return the data in the redis server
Else, do a web request and store the data in the redis server
You'll have to do this manually for the pages that you want while you have the internet connection. Once you've got all the data you need, you can work without the internet connection too.
If the pages are essentially static then you could use something like HTTrack http://www.httrack.com/ to make an offline copy
If there's anything requiring server side interaction or dynamic generation of pages you're most likely going to need to run your own local instance of the server.

Is there a google API to read cached content? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I know you can go to http://webcache.googleusercontent.com/search?q=cache:http://example.com/ to view Google's cache of any URL, but do they provide an API to hit thousands of these and pay for access?
I don't want to just make HTTP GETs to these URLs too fast and get my IP addresses banned or upset Google.
Just wondering if they offer a way to pay and do this through official channels like they do with their search API.
Google doesn't seem to have an API to access the cached results:
There are some attempts to scrape it and wrap it in APIs, such as this perl module
Other than that the Wayback Machine has an API, of cached versions of sites. Perhaps that will do?
Currently there's no tool that I've found that does it. You'd have to create your own script to individually cache a certain number of pages. To avoid Google blocking you, I suggest capping the number of urls scraped. Not ideal, but running a script 10 times is better than looking at 1000 cached urls individually. :/
If you want to see if anything you edit on your site would effect your potential rankings in Google, I'd check out SEORadar.com, they'll do that for you.

What service do you use to distribute software? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 4 years ago.
Improve this question
I work for a medium sized software company and have been put to the task of finding a new way of electronically distributing our software. We don't have a super fast connection to distribute it ourselves so it would need to be a solution that we can upload to and send out links to customers. The customers won't be purchasing our software from our website as we already do most of our sales from direct sales and partner sales. Since I joined the company we have grown from CD distribution sized downloads to DVD sized distribution downloads. We released a new version and find the YouSendIT Service to be clunky and 99% of our customers receive a link to download the software. We only send out a printed media if requested. Is there a service besides yousendit that allows for unlimited file size uploads/downloads. I have heard of drop.io and it seemed to be similar to yousendit. If you could please point me in the direction of Electronic software distribution system that is 3rd party hosted would be appreciated.
Thanks
Mike
You should look into Content Delivery Networks, such as Amazon CloudFront.
You might want to reconsider the way you are going about this.
If you software is open source, you should be using sourceforge. Otherwise you should just get a cheap hosting plan with lots of transfer bandwidth.
For example, godaddy has an unlimited account (unlimited transfer, unlimited space) for about $14.95 per month.
You point a sub domain i.e. download.rivageek.com to that server. This gives your users confidence when they download your application.
If they have to go to some ad laden 3rd party site they might think twice about giving you money. If you lose only 1 customer to that, it pays for itself (assuming you charge more than 14.95 for your product).
The fine print on many of those 3rd party sites mean they own whatever you upload as well.
If you'd like something that allows (simplisticly) secure one-time downloads, I've used filehosting.org in the past. They give you a hashed link to the software when you upload it, which you can then email to anybody you want to be able to download the file. If you want, you can set it to delete the file after one download.
In response to using your own domain for the downloads, it's possible to configure both Amazon S3 and CloudFront to use a custom domain name. Here are the instructions for S3 -- very straight forward:
http://docs.amazonwebservices.com/AmazonS3/latest/index.html?VirtualHosting.html
If emailing out a direct link to your distribution file (zip, etc.) is sufficient, I'd say go with one of these services -- they're very cost effective, reliable, and easy to set up.
You could use a filehosting service or get a regular web host with unlimited bandwidth just avoid Godaddy as its shared hosting is overcrowded and overbooked. (personal experience)

blocking website via hosts file not working? [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I'm trying to block a website using the hosts file using this tutorial : http://hackspc.com/how-to-block-a-website/
but It doesn't work, the website I'v blocked In this case facebook still shows up, please can anyone help me out here?
link textI could not access the link (blocked in office) but i think this may help you edit your host file
"“WWW” has become the universal standard for the default host. It is just as common to define a site with no host as well. This means that as far as DNS goes www.yahoo.com & yahoo.com are two totally different sites, even though they resolve to the same place. Because of this to effectively block the site, you must also block all hosts. This would usually mean:
127.0.0.1 facebook.com
127.0.0.1 www.facebook.com"
check the link for more details.
Also i you trying to block multiple sites, its better to have a proper software like proxy server or firewalls which can block access to particular sites.
I think Kavitesh Singh made the most important point: Blocking the domain with and without www. this is the most common reason for an entry not working.
Also, not all browsers immediately react to changes in the hosts file. Have you tried re-starting your browser and / or system?

Resources