We have upgraded to ColdFusion 10 and I am testing large upload capability.
Using both a HTML form and the flash multi-file upload CFFILEUPLOAD I can upload files of up to 2GB.
With files over 2gb the upload does not even start. 0% both with the flash upload and what chrome browser reports with HTML form.
Technical services suggest it does not even get as far as Apache, that is not restricting the upload. ColdFusion is also setup to allow 4000MB post data even with throttle.
The upload is occurring across the network, so even with test a 1.7gb file it doesn't take long - but 2.5gb does not even begin.
Any suggestions to help diagnose the cause?
Thanks
Related
I'm using the latest version of rclone to copy lots of files to backblaze b2 storage. What I've noticed recently is that there is a lot more download bandwidth being used than upload bandwidth. Eg running iftop on the network interface being used these are typical figures:-
Upload 77.6KB/s
Download 1.32MB/s
Why is there such so much being downloaded if all I am doing is backing up files to the server?
What can I do lessen the download bandwidth?
Many thanks
Have you tried the -bwlimit option to reduce download bandwidth?
I have configured my drupal site so that all images/files/media etc is handled my s3 by using S3 file system module.
Now everything works fine, the image/file/ field uploader works fine but there is a huge performance issue when using IMCE file browser from the WYSIWYG editor. It takes at least a minute for the browser to display its content and there are only 290 images with 78 MB used in that initial folder which should not cause such huge delays. This is having a huge impact for our editors and several minutes lost just to upload a couple of images.
I tried various pagination patch and there is no difference at all in the performance.
What are my options now
As drilled through many forms and discussions, turns out that IMCE was not meant for S3 file system and I found this patch in pdf form(warning downloads rather than opens )
I followed the steps in that patch which significantly improved my performance.
I'm hosting an HLS stream with XAMPP / Apache, which basically means I have a folder in my document root that contains a couple of incrementally numbered 10-second video files.
Every 10 seconds, a new video file is saved into the folder and the oldest video file in the folder is deleted.
Apart from these video files, the document root also contains some other files, such as PHP scripts and playlist files.
My server has plenty of RAM and a pretty fast CPU, but is using a comparatively slow hard disk.
Given the fact that the constant downloading of these video files is likely what's going to make or break the server performance, it seems like a good idea to cache these files in memory.
If Apache were to keep all video files (with a .ts extension) that're downloaded by a user's video player, in it's memory for about 60 seconds, the next user would then be able to download the file much faster. Apache could rely on the files not changing after the first open and on the fact that the files won't be requested anymore after those 60 seconds.
All other files do not (necessarily) have to be cached, since they're rather small and are regularly modified.
Is anyone able to give me directions on how to get started?
Modern operating systems already cache accessed files in memory. The whole process is managed by the kernel automatically.
Apache in-memory caching won't help you since it needs all the files at start-up.
If you want some level of control over the caching you could use vmtouch. Check the manual.
For our ecommerce store, we have a backend that allows employees to upload product shots. These images are then resized to fit the design of the website.
For a while now we've been having an issue with some very weird distortion in the images. Here is a sample of the distortion:
Reuploading the image usually fixes the problem, but it's been getting worse and worse.
Today, I took at look at the original uploaded files, and I see that the distortion is there in the original file. I know that they are not being uploaded like that, so the corruption is happening during the upload. Any idea what might cause this?
The originals are produced in Photoshop on Mac, and uploaded using Firefox on a Mac.
Here is the code that uploads the file
<cffile action="UPLOAD"
filefield="form.imageFile"
destination="#UploadPath#"
nameconflict="makeunique"
accept="image/gif, image/pjpeg, image/jpeg, image/jpg, application/pdf">
So as it turns out, the issue is was specific to that computer and that browser. Seems that that specific Firefox was corrupting files randomly on upload.
Using a different browser, even on that same machine, has resolved all the issues.
How can make a uploader file in vb6 to work easyly and can be uploade files into host.
For example (images or multimedia files or other format) and size > small size files for exampe > 500kb or 1 mb , 2 mb or upper sizes .
Please help me to can find any way to designed without ftp protocols.
I think it is possible to use of msxml refrences but i don't know how can designed it by msxml .
[thanks]
You could upload to a HTTP like dropbox. The downside of simple HTTP file transfer is that it is fairly easy for somebody to grab your credentials using a common tool like Fiddler. However that is also the case with FTP just a bit harder. The upside to uploading to dropbox is that you could have any file size (providing you have the space on your dropbox account).
I haven't tried using the dropbox upload api yet but you can read the documentation here, take a look at the PUT and POST methods: https://www.dropbox.com/developers/core/docs
If you just wanted to upload to dropbox and you have the dropbox client installed you could just save the file to the dropbox folder on your computer and it would auto upload.