Big File Upload Configuration

System Configuration

  • Make sure that the latest version of PHP, supported by ownCloud, is installed.

  • Disable user quotas, which makes them unlimited.

  • Your temp file or partition has to be big enough to hold multiple parallel uploads from multiple users; e.g. if the max upload size is 10GB and the average number of users uploading at the same time is 100: temp space has to hold at least 10 x 100GB.

In Centos and RHEL, Apache has a few more default configurations within systemd. You will have to set the temp directory in two places:

  1. In php.ini, e.g., sys_temp_dir = "/scratch/tmp"

  2. In /usr/lib/systemd/system/httpd.service:


Configuring Your Web Server

ownCloud comes with its own owncloud/.htaccess file. Because php-fpm can’t read PHP settings in .htaccess these settings must be set in the owncloud/.user.ini file.

Set the following two parameters inside the corresponding php.ini file (see the Loaded Configuration File section of PHP Version and Information to find your relevant php.ini files):

php_value upload_max_filesize = 16G
php_value post_max_size = 16G

Adjust these values for your needs. If you see PHP timeouts in your logfiles, increase the timeout values, which are in seconds, as in the example below:

php_value max_input_time 3600
php_value max_execution_time 3600


1707-update-big-file-upload-docs The mod_reqtimeout Apache module could also stop large uploads from completing. If you’re using this module and getting failed uploads of large files, either disable it in your Apache config or raise the configured RequestReadTimeout timeouts.

Disable mod_reqtimeout On Ubuntu

On Ubuntu, you can disable the module by running the following command:

a2dismod reqtimeout

Disable mod_reqtimeout On CentOS

On CentOS, comment out the following line in /etc/httpd/conf/httpd.conf:

LoadModule reqtimeout_module modules/

When you have done run asdismod or updated /etc/httpd/conf/httpd.conf, restart Apache.

There are also several other configuration options in your web server config which could prevent the upload of larger files. Please see your web server’s manual, for how to configure those values correctly:

Apache with mod_fcgid

If you are using Apache/2.4 with mod_fcgid, as of February/March 2016, FcgidMaxRequestInMem still needs to be significantly increased from its default value to avoid the occurrence of segmentation faults when uploading big files. This is not a regular setting but serves as a workaround for Apache with mod_fcgid bug #51747.

Setting FcgidMaxRequestInMem significantly higher than usual may no longer be necessary, once bug #51747 is fixed.

Configuring PHP

If you don’t want to use the ownCloud .htaccess or .user.ini file, you may configure PHP instead. Make sure to comment out any lines .htaccess about upload size, if you entered any.

If you are running ownCloud on a 32-bit system, any open_basedir directive in your php.ini file needs to be commented out.

Set the following two parameters inside php.ini, using your own desired file size values, as in the following example:

upload_max_filesize = 16G
post_max_size = 16G

Tell PHP which temp file you want it to use:

upload_tmp_dir = /var/big_temp_file/

Output Buffering must be turned off in .htaccess or .user.ini or php.ini, or PHP will return memory-related errors:

output_buffering = 0

Configuring ownCloud

As an alternative to the upload_tmp_dir of PHP (e.g., if you don’t have access to your php.ini) you can also configure a temporary location for uploaded files by using the tempdirectory setting in your config.php.

If you have configured the session_lifetime setting in your config.php. See Sample Config PHP Parameters, to make sure it is not too low. This setting needs to be configured to at least the time (in seconds) that the longest upload will take. If unsure, remove this entirely from your configuration to reset it to the default shown in the config.sample.php.

General Upload Issues

Various environmental factors could cause a restriction of the upload size. Examples are:

  • The LVE Manager of CloudLinux which sets an I/O limit.

  • Some services like Cloudflare are also known to cause uploading issues.

  • Upload limits enforced by proxies used by your clients.

  • Other webserver modules like described in General Troubleshooting.

Long Running Uploads

For very long running uploads (those lasting longer than 1 hr) to public folders (and chunking is not in effect), 'filelocking.ttl' should be set to a significantly large value. If this is not set appropriately, then large file uploads will likely fail. A simple way to calculate the correct value is by using the formula: time in seconds = (maximum file size / slowest assumed connection).