Search Results

Search found 17278 results on 692 pages for 'directory conventions'.

Page 377/692 | < Previous Page | 373 374 375 376 377 378 379 380 381 382 383 384  | Next Page >

  • turn off disable the performance cache

    - by jessie
    OK I run a streaming website and my CMS is giving me an error when uploading videos "Failed To Find Flength File" ok so I did some research. The answer I got from the coder was below. I did do all that, but the only thing I could not do is turn off what he refers to as performance cache, talked about in the last sentence... I am on a Cent OS Assuming the script is set up properly, you are probably dealing with some kind of write-caching. Some servers perform write-caching which prevents writing out the flength file or the entire CGITemp file during the upload. The flength file or the CGITemp file do not actually hit the disk until the upload is complete, making it worthless for reporting on progress during the upload. This may be fixed using a .htaccess file assuming your host supports them. Here is a link to an excellent tutorial on using .htaccess files. I strongly recommend giving it a quick read before attempting to install your own .htaccess file. 1. A mod_security module for Apache. To fix it just create a file called .htaccess (that's a period followed by "htaccess") and put the following lines in that file. Upload the file into the directory where the Uber-Uploader CGI ".pl" scripts resides, or in some directory above it (like your server's DOCUMENT_ROOT, i.e. the top-level of your webspace). htaccess files must be uploaded as ASCII mode, not BINARY. You may need to CHMOD the htaccess file to 644 or (RW-R--R--). # Turn off mod_security filtering. SecFilterEngine Off # The below probably isn't needed, # but better safe than sorry. SecFilterScanPOST Off If the above method does not work, try putting the following lines into the file SetEnvIfNoCase Content-Type \ "^multipart/form-data;" "MODSEC_NOPOSTBUFFERING=Do not buffer file uploads" mod_gzip_on No 2. "Performance Cache" enabled on OS X SERVER. If you're running OS X Server and the progress bar isn't working, it could be because of "performance caching." Apparently if ANY of your hosted sites are using performance caching, then by default, all sites (domains) will attempt to. The fix then is to disable the performance cache on all hosted sites.

    Read the article

  • Using relative path in PHP.INI doc_root -- getting No Input File Specified -- running as fastCGI

    - by J.R.
    I'm attempting to run php-cgi under LightTPD on Windows with my doc_root set to one directory up (doc_root = "../Docs") -- however, I get "No input file specified". I've set cgi.fix_pathinfo, and all the other tricks I could find with no success. If I set doc_root to an absolute path, it works fine. How can I make this work? If any additional information is required, I'll gladly provide it. Thanks in advance.

    Read the article

  • IIS, SSL with client certs on web farm

    - by Jeremy
    We're building a web service that will be deployed on an IIS 7.5 farm, and secured through SSL, and also requiring client certs that will be mapped to Active Directory accounts. My understanding is that the server cert needs to be generated for a specific server. If that is the case then we will need a server cert for each server in the farm. Because the farm will be load balanced, how do we generate client certs that will work with any of the servers in the farm?

    Read the article

  • Apache configuration FollowSymlinks- Apply to php scripts?

    - by Josh
    I have Options set to none for my webroot directory. I also have a symmlink /var/coderoot - /var/webroot/coderoot In the php script I can do include("/var/coderoot/file"); and it works fine. Regardless of the option (yes I save and restart apache.) Does follow symlinks only apply to symlinks used in a certain way? Is there a performance loss using the include with a symmlink?

    Read the article

  • Automatically sync directories via FTP during off hours?

    - by jason
    I would like COMPUTER_A to sync anything found in a specific directory, with my FTP server, COMPUTER_B, but only during off hours. I would like it to automatically resume if the computer is rebooted. I use FileZilla now, but it will not automatically continue transferring when the computer is restarted. I also do not think that you can set the times for the transfers, with FileZilla.

    Read the article

  • Are PHP session files ever deleted?

    - by GetFree
    I see there are thousands of files in my "/tmp" directory (a CentOS machine) and almost all of them are PHP session files. I'm worried about the possible impact this might have on my system. Are those files ever deleted either by the OS, Apache or PHP? or I have to take care of it myself?

    Read the article

  • Linux: Create files and direcotires but not delete them

    - by Peraz
    I have a process that create directories and files inside a working directory, ex: /workingdir/file1 /workingdir/file2 /workingdir/dir1/file1 /workingdir/dir1/dir2/file1 /workingdir/dir1/file2 I need to avoid deletion/overwrites of created folders/files for that user, but allow subsequent folders/subfolders/files creation. I try permissions, gid, acl with no luck. What is the correct way to do that ? (i can use a cron job if needed)

    Read the article

  • Lazy linux backup system?

    - by Alex
    Ok, so I want to say Time Machine, but that's note exactly what I'm looking for. I want to set up a system that will regularly (hourly?) back up the /home directory of our machine. time Machine style things are naturally preferable since they save space by only saving the changes, but honestly, this is important enough that I can suffer some waste. Any ideas?

    Read the article

  • unable to connect site to different port

    - by JohnMerlino
    I have a domain was registered at godaddy named http://mysite.com/. I logged into godaddy and I went to All Products Domains Domain Management. I clicked on the appropriate domain and it took me to the Domain Details page. I clicked Launch under DNS Manager and it took me to the Zone File Editor. I noticed that notify.mysite.com was pointing to an IP address pointing to a dead server, so I switched it to an operating server. Then I pinged the domain to see where it was pointing to and it was correctly pointing to the working server. So I copied the default configuration under sites-available: sudo cp default notify.mysite.com. And then I made some edits to it to have it point to a different document root to serve files at a different port: Listen 1740 Listen 64.135.xx.xxx:1740 #I also tried this as well: NameVirtualHost 64.135.xx.xxx:1740 <VirtualHost 64.135.xx.xxx:1740> ServerAdmin [email protected] ServerName notify.mysite.com DocumentRoot /var/www/test/public <Directory /var/www/test/public> Order allow,deny allow from all </Directory> ErrorLog ${APACHE_LOG_DIR}/error.log LogLevel warn CustomLog ${APACHE_LOG_DIR}/access.log combined </VirtualHost> Then I enabled the virtual host. Then I went to the document root and added an index.html file with some text in it. Then I restarted apache. The restart gave no errors. Then I type the correct domain in URL: http://notify.mysite.com:1740/ and I get: Oops! Google Chrome could not connect to notify.mysite.com:1740 Somehow it took out all my other sites. Now even the ones that were responding on port 80 are no longe responding, even though I did not touch the virtual hosts for them. I get this message now: Oops! Google Chrome could not connect to mysite.com However, ping responds: ping mysite.com PING mysite.com (64.135.12.134): 56 data bytes 64 bytes from 64.135.12.134: icmp_seq=0 ttl=49 time=20.839 ms 64 bytes from 64.135.12.134: icmp_seq=1 ttl=49 time=20.489 ms The result of telnet: $ telnet guarddoggps.com 80 Trying 64.135.12.134... telnet: connect to address 64.135.12.134: Connection refused telnet: Unable to connect to remote host

    Read the article

  • setting up a samba PDC -error with testparm

    - by Rungano
    Hi guys I have installed a samba PDC but when I test the samba configurations file I am getting errors like these, "Invalid combination of parameters for service homes. Map system can only work if create mask includes octal 010 (S_IXGRP)." My Configuration file is as follows [homes] comment = Home Directories path = /home_srv1/%u valid users = %S read only = No create mask = 0660 directory mask = 0770 browseable = No I tried to google but with no luck, Serverfault is always my best hope. Thanks for helping out.

    Read the article

  • cygwin, PATH problem?

    - by jayjaypg22
    I run a .ksh containing a awk call. awk.exe and his shortcut awk is in /bin/awk, /bin is in the PATH environment variable. But when I try to launch awk, I have this error message : bash: /usr/bin/awk: no such file or directory Why didn't bash look for it in the /bin folder too? edit : tar has the same rights, tar.exe is in /bin and can be listed in /usr/bin/, the exact same way than awk. Tar works fine whereas awk not.

    Read the article

  • Can't add a .ldif to OpenLDAP, recent version (no slapd.conf)

    - by Biganon
    I'm new to LDAP and I'm trying to add the mmc.ldif and mail.ldif files that come with Mandriva Directory Server to my LDAP configuration, using the command : ldapadd -x -W -D "cn=admin,dc=biganon,dc=com" -f schema/mmc.ldif I then give the admin password I've set during slapd installation but get this error : ldap_bind: Invalid credentials (49) I have no slapd.conf file. I'm on Debian 6.0.1, OpenLDAP 2.4.23 Thank you

    Read the article

  • Windows 7 Explorer shortcut for "Replace All"

    - by chris
    I you copy/paste some files into a directory that already contains files with the same name you'll get a confirmation dialog "There is already a file with the same name in this location..." To replace all files you have to check "Do this for the next N conflicts" and then click on "Copy and Replace". Is there a keyboard shortcut for this (like there was in XP where you could simply press 'a')?

    Read the article

  • how do I copy from a remote linux machine to a local linux machine?

    - by leonbnu
    I sshed to a remote machine and want to copy some files from there to my local machine. Both are running linux. My local machine is behind a firewall, so I can't do "scp file_on_remote user@localmachine". I know I can do "scp user@remotemachine:/directory/file ./", but I need to wander around on the remote server to pick some files to download to local machine, so this way doesn't seem too appealing to me.

    Read the article

  • What is the CPU microcode?

    - by golem
    Today after unpacking the initramfs file from my Linux /boot directory I found that it contains the GenuineIntel.bin file, and that that file is an image of what is called the CPU microcode. Then I found articles on ArchWiki and Wikipedia describing the subject. Now I'm not sure that my vague understanding of the subject is correct. Can anyone please explain what the CPU microcode is in plain English?

    Read the article

  • Prevent Xbox users from editing Media Library

    - by Patrick
    I am trying to watch some videos stored on my desktop computer on the Xbox, but they are in a format that the Xbox cannot decode, so I have to stream them through Windows Media Center. However, as soon as I set up Media Center on the Xbox, anyone can go in and browse the directory structure on my desktop. I would like to "lock" the media library so that only I can add and delete folders from my desktop computer. Is it possible?

    Read the article

  • Can I copy a cross compiler tool chain between systems (I did before)?

    - by Jamie
    I tested fairly extensively with Ubuntu 10.04 Beta 2 Server in a VM, and was able to simply copy (read tar x) a cross compiled tool chain from an Ubuntu 8.10 VM. I created the tar myself, which is essentially a lot of stuff in \usr\local. Now that I've got a bare metal installation of Ubuntu 10.04 proper, the copy isn't working. In particularly, I'm getting the error: $ arm-linux-gcc -bash: /usr/local/bin/arm-linux-gcc: No such file or directory I've got the systems side by side in SSH windows ... any suggestions?

    Read the article

  • how to redirect itunes to external hard drive

    - by oo
    i have all my music on my hard drive but its running out of space. i bought an external hard drive, copied all my music over i went into itunes advanced preferences and changed my music location but when i view the itunes music library XML file in notepad, i still see a lot of things pointing to my regular hard drive i didn't check "Keep music files organized" because i didn't want itunes trying to copy everything over from my regular hard drive (because i already did it) is there any other way to simply refresh the music list given a new root directory.

    Read the article

  • Execute a program by double-clicking the taskbar and/or the desktop, Windows 7

    - by JoePerkins
    It would be so useful, and not very difficult, because in fact I have Directory Opus installed (a very powerful windows explorer alternative) and it does exactly that, but only by double-clicking the desktop, not the taskbar. Similar options with middle-click would also be nice, such as scrolling the taskbar (maybe cycling directly through the Alt-Tab window). 7 Task Tweaker is close to this, actually, but it doesn't do what I would like.

    Read the article

  • why is nautilus slow?

    - by Coding District
    I'm wondering why Nautilus is very slow when opening a directory containing lots of files. My /usr/lib dir for example has 1900 files and it takes approximately 5+ seconds to show everything. It has been like this since I installed Ubuntu few months ago and it's really quite annoying sometimes. I don't have powerful hardware but I know that Windows Explorer is so much faster than this. Is there anything that can be done to speed it up? Ubuntu 10.04

    Read the article

< Previous Page | 373 374 375 376 377 378 379 380 381 382 383 384  | Next Page >