Search Results

Search found 21717 results on 869 pages for 'setup versions'.

Page 542/869 | < Previous Page | 538 539 540 541 542 543 544 545 546 547 548 549  | Next Page >

  • Why can I not access any file or directory created by PHP from FTP-client?

    - by user43053
    Hello there, If I create a directory with mkdir(), or create a file with fopen(), file_put_contents() or SimpleXMLElement::asXML(), I am unable to access the file with my FTP-client or c-Panel File Manager. If I try to delete or edit them, I get errors. Dreamweaver suggests it is a permission problem or a network or filesystem fault (but I've set the permissions with chmod() to 0777, and when I check the cPanel, it confirms chmod 777. I also tried to use fileowner() and the function returns int(99), the same owner as those files that I could access with my FTP-client. It seems files and directories created with PHP can only be modified or be deleted with PHP. I thought this must be a server setup related issue, so I write it here. I am on a shared server, and I have no idea about setting up servers. Thank you for your time. Kind regards Marius

    Read the article

  • Making my SVN Public

    - by azz0r
    Hello, I'm looking todo an SVN checkout on a server so I need to make my local SVN public. I looked into GITHUB, but I'm not willing to pay or let the world see my project. Are there any alternates? Okay so I went through this tutorial: http://www.petri.co.il/setup-ssh-server-vista.htm Had some issues, so I did this: mail-archive.com/[email protected]/msg84875.html Now I'm wondering how let the SSH access my SVN repo found in c:/wamp/svnRepo. Any tutorials or advice (please no: go read this book crap) greatly welcome!

    Read the article

  • I Can't run the Netbeans but I installed successfully

    - by David
    I'm new to Ubuntu as well as Netbeans. I installed Netbeans, and I've made sure to install all the JDKs and JREs I could find. It installed without errors. I also saw this question and made sure I followed all the instructions there as well. I never got any error messages of any kind. So far as I know, it installed okay. However, when I try to run Netbeans, I get the message in the bottom of the Netbeans IDE like this: ant -f /root/NetBeansProjects/samp1 -Djsp.includes=/root/NetBeansProjects/samp1/build/web/one.jsp -DforceRedeploy=false -Dclient.urlPart=/one.jsp -Ddirectory.deployment.supported=true -Djavac.jsp.includes=org/apache/jsp/one_jsp.java -Dnb.wait.for.caches=true run /root/NetBeansProjects/samp1/nbproject/build-impl.xml:774: The libs.CopyLibs.classpath property is not set up. This property must point to org-netbeans-modules-java-j2seproject-copylibstask.jar file which is part of NetBeans IDE installation and is usually located at <netbeans_installation>/java<version>/ant/extra folder. Either open the project in the IDE and make sure CopyLibs library exists or setup the property manually. For example like this: ant -Dlibs.CopyLibs.classpath=a/path/to/org-netbeans-modules-java-j2seproject-copylibstask.jar BUILD FAILED (total time: 0 seconds)

    Read the article

  • How to identify which website on my instance is receiving lots of traffic?

    - by Bob Flemming
    I am new to server administration and have just setup a new quad core instance which hosts around 15 websites. Over the past couple of days my server load has been averaging at around 15.00. I believe it is because of one (or maybe more) websites are getting spammed by spambots. Typing 'top' at the command line shows many processes from user 'www-data' which indicates lots of web traffic. Is there an easy way identify which one of my sites is taking a hammering? Reading the apache error logs is a very difficult tasks as most of the websites receive daily traffic of 10,000 + unique users. Any help would be appreciated!

    Read the article

  • How to Use Vim-Style Keyboard Shortcuts for OS X Tab Navigation

    - by The Geek
    After switching to OS X when I got a new MacBook Air, one of the first things I needed to duplicate was my extremely customized AutoHotkey setup — the most important of which is using the J and K keys to navigate throughout tabbed windows easily. Yeah, I’m a Vim user. I’ve never been a fan of having to use CTRL + TAB to switch from one tab to the next — to start with, you have to move your hands from the home row, and it’s awkward, and why should I have to do that just because somebody decided that keyboard shortcut before tabs became popular? If you think about it, if tabbed browsers were popular back when keyboard shortcuts were being invented, they would have definitely reserved some of the good shortcuts for switching tabs. On Windows, I’ve always used an AutoHotkey script to make things the way I wanted it:  ALT + J and ALT + K for selecting previous and next tabs. Once you get used to it, it’s extremely awesome, and so much faster than using CTRL + TAB. Of course, I also hacked CTRL + T and CTRL + W into ALT + T and ALT + W so I could open new tabs and close them without moving my hands from the home row. Over on OS X, it turns out that it’s incredibly simple and easy to use CMD + J and CMD + K for next/previous tab navigation, and it works in most applications that support tabs, like Terminal, Safari, or Google Chrome.    

    Read the article

  • Ubuntu server 11 showing weird characters on console

    - by andySF
    I've installed Ubuntu server 11 after my last hard drive failed to perform. The install process and GRUB screen are OK but I have a problem with the console. I attached two photos because it's difficult for me to describe this issue. I have little knowledge about Linux. My server is used for git, file server, print server. I played with this command dpkg-reconfigure console-setup but no effect. It's nothing physicality wrong with my display or my memory(it's on board). I have run a memtest, and besides that withe the previous version of Ubuntu I've had no problems. The problem.

    Read the article

  • Custom PC won't boot Windows 7 dvd but does with windows vista

    - by M_rk
    I ordered a custom build PC. (It was assembled by the store). This is the setup Motherboard: Asrock A75 PRO4-M DVD drive: LG GH24NS90 SSD: Samsung 830 series 128GB DDR3 SDRAM: 2× Corsair XMS3 CMX4GX3M1A1600C9 (2× 4GB) APU (CPU+GPU): AMD A8-3850 Boxed I got a installation DVD for Windows 7 Professional x64 English (including SP1), but it doesn't work. I got a new one from the store and it doesn't work either. However they work on a other PC. So the DVDs aren't bad. I tried an old installation DVD for Windows Vista. Both 32 bit and 64 bit work. So the boot order and such are right and working on the new PC. Is there something I'm missing here? Any ideas on how to make it work?

    Read the article

  • How to test Laptop NIC's throughput using a router and PC - without be bounded?

    - by 0x90
    My setup includes: Cisco router An i-7 PC running windows A laptop with high speed wifi nic, which I want to check its throughput. I would like to run an FTP server on the PC. hook the router over cables to the PC. I would like to have the PC create its own subnet accessible via the cisco router that would be hooked directly to the PC's nic. From the laptop I want to connect via wifi to the PC's wireless router and connect to the ftp server on the PC. is it possible? how do i connect the router to the PC nic and make it broadcast a subnet via wifi for my laptop to connect to? how do i configure an FTP server to operate only on this subnet?

    Read the article

  • Interaction between two Clouds

    - by user7969
    I have setup the Cloud-A with 1 - [CLC+CC] and 2 - [NC] computers. I have another Cloud-B with same configuration. [using the Ubuntu Enterprise Cloud] Both of them working fine individually, in the same LAN. Now if I want to add the NC of Cloud-A to CC of Cloud-B, [in case the resources of Cloud-B are exhausted] how can I make it possible ? I guess this calls for the interoperability stuff... Could you please explain what happens exactly when we ask for instance, the direct interaction happens between the client and NC or it goes through the CLC and CC ? What I want to say is, say there are multiple cloud providers. A user is subscribed to any one of them, say Cloud-A for IaaS. As the requirements are dynamic, all the resources of Cloud-A may get exhausted. There may be another Cloud-B which can provide the services but that Cloud-A can't ask the client to go for Cloud-B. So if it is possible to have some co-ordination between this two providers to share resources mutually, making client fully unaware of whats going on in the background....? Please reply.. I am sorry if I'm doing mistake anywhere... Thanks in advance :)

    Read the article

  • Installing additional printer drivers (x86) on Windows 7 (x64)

    - by ambrax
    I've successfully installed Windows 7 (x64) and drivers for my Canon MP510, and have no problem printing with this setup. There is another PC in the network running XP SP3, and I want to share the printer so that users of that PC can also print. On W7 I have the option of installing additional printer drivers for other system architectures (Itanium and x86). I've downloaded the most current 32-bit drivers for the printer, but every time I direct the install dialog to the folder containing the drivers, I get the following error message: Selected printer driver not found The specified location does not contain the driver Canon Inkjet MP510 Printer for the requested processor architecture. Retry Cancel I'm stumped. I'm absolutely certain that the specified location actually does contain the correct drivers; I've even installed them on the XP system. I've tried everything I can think of. What am I overlooking?

    Read the article

  • Booting Ubuntu 12.04 from external eSATA disk

    - by Lord of Scripts
    This is my system topology: Disk #1 (SATA Internal) C: D: (Windows 7 Ultimate) Disk #2 (SATA Internal) E: (Windows Backup) Disk #3 (eSATA External) H: I: (Other windows data) /dev/sdc3 Linux Swap /dev/sdc4 Extended partition /dev/sdc5 Linux / So, I originally had there Ubuntu 8.1 from years ago but never got to use it. Now I used the Ubuntu 12.04 Live CD to install on that same location (That live CD takes a century to boot on a 6GB Intel i7 system...). The installation went fine, I selected it to install on /dev/sdc5 but it never asked me for any boot stuff, where I wanted to install Grub or whatever it is that it uses nowaways (I come from the LILO days when it always worked :-) So, yet again I can't access my new Linux installation. I have to wait a century to boot the "Live" CD and it allows me to see my new installation but I can't do anything with it. I tried the approach of this blog post. Copied the linux.bin of /dev/sdc5 into C: and used the BCDEdit steps to declare the new OS. So when I boot I see the Windows Boot menu and select Linux and after than I only get a black screen with a blinking cursor on the upper left. I can boot into Windows though. So, perhaps it didn't install the boot code on /dev/sdc5? I used this setup years ago booting from Windows with a BIN file: dd if=/dev/sdc5 of=/mnt/share/C/linux.bin bs=512 count=1 I am very reluctant to run GRUB because years ago I did and it wiped out my Windows boot sector and took quite some effort to recover it and be able to boot Windows again. I have been trying to install GRUB on a blank USB stick but I can't find anything clear enough. My system does NOT have a floppy. So can someone give me some ideas about how to get control of my Ubuntu 12.04 installation?

    Read the article

  • Ownership/permissions of uploaded files

    - by Cudos
    Hello. I want to find out if I am on the right track. My script uploads files to the directory "images". The directory has this setup: owner/group = www-data Permissions = 700 Questions: Is this a good way to secure the directory from a hacker uploading files? Will the hacker be able to upload the files directly to the directory? Note: I have a bunch of other security measures in my upload script + an .htaccess script in the directory that disables script Execution. I just what to know if the permissions on the directory is sensible. I run apache 2.2

    Read the article

  • Dropbox context menu missing in OS X

    - by slhck
    Problem My Dropbox context menu is missing in OS X Snow Leopard (10.6.8). While the Dropbox service runs normally, Finder doesn't show the icons and also doesn't give me the ability to browse files on the website or copy the public link. What I've tried Removed ~/.dropbox and ~/Dropbox/.dropbox.cache Reinstalled Dropbox.app (both 1.4.7 stable and 1.5.0 experimental), went through the setup again Restarted Finder Logged out and back in All of these I've done over and over again, in random permutations. I've made sure that Dropbox appears in the Login Items under my Account (and I've never touched that) I don't know if ~/Library/Contextual Menu Items is missing the Dropbox plugin or if there shouldn't be one after all. In any case, I can't get the icons or the menu to appear.

    Read the article

  • Using pkexec policy to run out of /opt/

    - by liberavia
    I still try to make it possible to run my app with root priveleges. Therefore I created two policies to run the application via pkexec (one for /usr/bin and one for /opt/extras... ) and added them to the setup.py: data_files=[('/usr/share/polkit-1/actions', ['data/com.ubuntu.pkexec.armorforge.policy']), ('/usr/share/polkit-1/actions', ['data/com.ubuntu.extras.pkexec.armorforge.policy']), ('/usr/bin/', ['data/armorforge-pkexec'])] ) additionally I added a startscript which uses pkexec for starting the application. It distinguishes between the two places and is used in the Exec-Statement of the desktopfile: #!/bin/sh if [ -f /opt/extras.ubuntu.com/armorforge/bin/armorforge ]; then pkexec "/opt/extras.ubuntu.com/armorforge/bin/armorforge" "$@" else pkexec `which armorforge` "$@" fi If I simply do a quickly package everything will work right. But if I package with extras option: quickly package --extras the Exec-statement will be exchanged. Even if I try to simulate the pkexec call via armorforge-pkexec It will aks for a password and then returns this: andre@andre-desktop:~/Entwicklung/Ubuntu/armorforge$ armorforge-pkexec (armorforge:10108): GLib-GIO-ERROR **: Settings schema 'org.gnome.desktop.interface' is not installed Trace/breakpoint trap (core dumped) So ok, I could not trick the opt-thing. How can I make sure, that my Application will run with root priveleges out of opt. I copied the way of using pkexec from synaptic. My application is for communicating with apparmor which currently has no dbus interface. Else I need to write into /etc/apparmor.d-folder. How should I deal with the opt-build which, as far as I understand, is required to submit my application to the ubuntu software center. Thanks for any hints and/or links :-)

    Read the article

  • check_ssh executed from Nagios server returns "Server answer: ", nothing else

    - by phs
    Commmand works on the host I need to monitor (its IP is denoted as $HOSTADDRESS$ here): [root@host ~]# /usr/local/nagios/libexec/check_ssh localhost SSH OK - OpenSSH_4.3 (protocol 2.0) However, on the Nagios server, it does not: root@server:~# /usr/lib/nagios/plugins/check_ssh -H $HOSTADDRESS$ Server answer: I can see that the port is open: root@server:~# /usr/lib/nagios/plugins/check_tcp -H $HOSTADDRESS$ -p 22 TCP OK - 0.123 second response time on port 22|time=0.123277s;;;0.000000;10.000000 And I can verify with telnet: root@server:/etc/nagios3/conf.d# telnet $HOSTADDRESS$ 22 Trying $HOSTADDRESS$... Connected to $HOSTADDRESS$. Escape character is '^]'. After searching the web for two hours and trying different suggestions I am out of ideas. I have several other hosts with similar setup, and they don't have this problem. I have compared the configs and they look the same (I think). I am clearly missing something here.

    Read the article

  • Change ownership of directory and all contents to a new user from root.

    - by Andrew Fashion
    I created a website under /var/www/html/ all under root, all images, files, .htacess, directories, etc... I uploaded and configured everything as root. I want to make it it's own username/password so it's not owned by root. I currently do not have the user account made either, I want to also setup FTP for the user account. There is also about 30GB of images in the folder as well. How can I go about changing all of this? I am running CentOS 5.5 64 bit. Thank you!

    Read the article

  • How do I get a 300Mbps connection over 802.11n?

    - by Daniel Schaffer
    I just bought a new wireless setup, consisting of a Cisco E2000 router, Edimax 7718un USB adapter for my laptop and an Edimax 7728in PCI adapter for my HTPC (which isn't in a location I can run cat5 to). I have to stay in the 2.4GHz band because I have an iPhone and a Wii that will need to connect. I'm aware that 11g devices will drop speeds for 11n devices, but they aren't connected yet. The fastest connection I've been able to get, with the router within 5 feet of either the laptop or HTPC is 144Mbps. The router has settings for "20MHz" and "Auto (20 MHz or 40Mhz) channel width, which I've set to the "Auto" setting. I haven't been able to find anything similar for either of the Edimax adapters. This is the first I've dealt with 11n, so I'm not even sure what else could be causing a problem. How do I get up to 300Mbps, or at least a fair bit closer?

    Read the article

  • Use HAProxy or Nginx to Load Balance between VPS

    - by xperator
    I want to load balance + failover backup multiple vps webservers hosted on different providers. I heard that for HAProxy you need multiple server under the same subnet, plus a shared (virtual) ip address between load balancers. But it's not possible in my case cause every VPS is on different node/network. Is there a way to use HAProxy in this kind of setup ? ( Please explain how briefly, I don't want to hear your "YES" answer ) What about NginX? Is it possible to achieve same result with Nginx ? (when servers are located on different nets) I know about Round Rubin DNS, but it doesn't provide a real failover solution, neither a load balance between servers.

    Read the article

  • Apache expands REQUEST_FILENAME with filetype

    - by Lauer
    I have a .htaccess fil on a Apache22 server. RewriteEngine On RewriteBase / RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.+)$ /index.php?p=$1 [L] The idea is when the request is /testpage it should be rewritten to /index.php?p=testpage. The problem is that apache does a write before my file, so the actually request is testpage.xml which actually does exists. Therefore, the first RewriteCond is newer true, and my rewrite rule is never applied on the /testpage request. There server should be a standard setup on a FreeBSD server. I know this work on other servers.

    Read the article

  • Why does Hyper-V and Windows Backup crash (BSOD) after successfull backup?

    - by Payson Welch
    Hello I am running Server 2008 R2 with a handful of Hyper-V guest nodes. If Windows backup runs without any of the Hyper-V nodes running, the server is fine. If Hyper-V runs a backup while the Hyper-V nodes are running, it is fine until a few minutes after the backup completes, and then it BSODs. The storage location for the backup is iSCSI - I am wondering if anyone has any input on what might be causing this? I don't have the Hyper-V nodes setup on a vlan and there is only one NIC on the server. Is it possible this is a networking / driver issue, and if so how would I reconfigure the networking to fix this?

    Read the article

  • WebLogic Partner Community Newsletter October 2013

    - by JuergenKress
    Dear WebLogic Partner Community member, Our October newsletter edition focuses on Oracle OpenWorld 2013, highlights, keynotes and all presentations. Thanks to all the partners who made the conference a huge success, if you could not come to San Francisco, you can find all the details in this newsletter. We added additional locations for the free hands-on ADF & ADF Mobile Bootcamps & WebLogic Bootcamps. As a community member you can also get a free voucher to become a WebLogic Server 12c Certified Implementation Specialist or ADF 11g Certified Implementation Specialist (limited to partners from EMEA!) If you can not make it to a Bootcamp, do not miss the virtual developer days for WebLogic and ADF Mobile. If you plan to install WebLogic read first the article “Setup a 12c Fusion Middleware Infrastructure from René van Wijk. If you administrate Middleware make sure you read the documentation and support notes Weblogic Server Patching & Maintenance Information Center. In the ADF section of the newsletter our product management team continues with the ADF Architecture on-demand training. Andrejus released the latest version of the ADF Performance Audit Tool v 2.0. The summer is over, if you look for a Christmas present, for your kids or yourself maybe you want to run Java on LEGO® Mindstorms® EV3. Jürgen Kress To read the newsletter please visit http://tinyurl.com/WebLogicNewsOctober2013 (OPN Account required) To become a member of the WebLogic Partner Community please register at http://www.oracle.com/partners/goto/wls-emea ( OPN account required). If you need support with your account please contact the Oracle Partner Business Center. Blog Twitter LinkedIn Mix Forum Wiki Technorati Tags: WebLogic Community newsletter,newsletter,WebLogic,WebLogic Community,Oracle,OPN,Jürgen Kress

    Read the article

  • public_html permissions for local development

    - by maGz
    I know this question has popped up a couple times, but I can't seem to find a definitive answer to my issue, so please bear with me. I have Ubuntu Server 12.04 setup in VirtualBox for PHP development and testing (Drupal plus other PHP sites using Yii framework). My question is in 3 parts... 1) If I create a public_html folder under /home/myuser, do I need to give ownership of that folder to the Apache www-data group? If so, are there any specific permissions I should be setting? 755? (Btw, I am following this guide to create the public_html directory and set up multiple virtual hosts per site I create and test) I previously had all of my sites under /var/www, but ran into massive permission denied errors whenever I tried to sFTP to it, either through FileZilla or PhpStorm. This is what I had previously done: sudo chgrp www-data /var/www sudo chmod -R 775 /var/www sudo chmod -R g+s /var/www sudo usermod -G www-data [my_ftp_user] 2) The second part of my question is this: If I create my PHP project and files in Windows through PhpStorm, and then upload via sFTP, will permissions get affected? 3) Once I am satisfied with my developed project, would it be advisable to move and test them under /var/www to see how it would fair in a production-ish environment? I would really appreciate the help and advice here. I'm learning more as I go along, but dealing with Linux files and permissions is a bit of a new ballgame for me! Thank you

    Read the article

  • PPTP: Linux clients unreachable

    - by Bahman
    I have setup a PPTP server on a CentOS 6 box. Several clients (Windows and Linux) connect to this server. All clients use the same subnet. All clients (Win or Lin) can surf the Internet using the VPN successfully. Windows clients can ping any other Windows client (XP/Vista/7) connected to VPN. Linux clients (KVPNC) can ping any Windows client (XP/Vista/7) connected to VPN. So far so good. But there's this weird problem and I've been trying to figure it out with no luck: No client can ping (reach) a Linux client (successfully) connected to VPN. And yes, the firewall is completely disabled. I'd really appreciate any hint/idea. TIA, EDIT: The ultimate goal of this is to share services across the VPN, like accessing a HTTPD running on a Linux box from a Windows client.

    Read the article

  • Difference between SSLCertificateFile and SSLCertificateChainFile?

    - by chrisjlee
    Normally with a virtual host an ssl is setup with the following directives: Listen 443 SSLCertificateFile /home/web/certs/domain1.public.crt SSLCertificateKeyFile /home/web/certs/domain1.private.key SSLCertificateChainFile /home/web/certs/domain1.intermediate.crt From: For enabling SSL for a single domain on a server with muliple vhosts, will this configuration work? What is the difference between SSLCertificateFile and SSLCertificateChainFile ? The client has purchased a CA key from GoDaddy. It looks like GoDaddy only provides a SSLCertificateFile (.crt file), and a SSLCertificateKeyFile (.key file) and not at SSLCertificateChainFile. Will my ssl still work without a SSLCertificateChainFile path specified ? Also, is there a canonical path where these files should be placed?

    Read the article

  • How do I stop someone from saturating my line & wasting CPU cycles

    - by JoshRibs
    My web host shows inbound & outbound traffic with mrtg. I have a steady 3.5mbps inbound traffic from Nigeria. Even assuming the source IPs & destination ports are blocked with Iptables & verifying nothing is listening on those ports, will the traffic still always pass through the switch & "get" to my server (where my server wastes CPU cycles "dropping" the packets)? Assuming I was setup with a hardware firewall, the traffic would still show in mrtg assuming the firewall is behind the switch? So is there any way to stop someone from saturating your 100mbps line, if they also have a 100mbps line? Other than filing an abuse complaint with the kind folks in Nigeria?

    Read the article

< Previous Page | 538 539 540 541 542 543 544 545 546 547 548 549  | Next Page >