Search Results

Search found 15439 results on 618 pages for 'wls configuration'.

Page 369/618 | < Previous Page | 365 366 367 368 369 370 371 372 373 374 375 376  | Next Page >

  • Troubleshooting VMware on Ubuntu

    Summary of different problems while using VMware products on Ubuntu. This article is going to be updated from time to time with new information about running VMware products more or less smoothly on Ubuntu. Following are links to existing articles: Running VMware Player on Linux (xubuntu Hardy Heron) Running VMware Server on Linux (version 1.0.6 on xubuntu) Using ext4 in VMware machine   VMware mouse grab/ungrab problem (Source: LinuxInsight) Upgrading GTK library in Ubuntu since Karmic Koala gives you a strange mouse behaviour. Even if you have "Grab when cursor enters window" option set, VMware won't grab your pointer when you move mouse into the VMware window. Also, if you use Ctrl-G to capture the pointer, VMware window will release it as soon as you move mouse around a little bit. Quite annoying behavior... Fortunately, there's a simple workaround that can fix things until VMware resolves incompatibilities with the new GTK library. VMware Workstation ships with many standard libraries including libgtk, so the only thing you need to do is to force it to use it's own versions. The simplest way to do that is to add the following line to the end of the /etc/vmware/bootstrap configuration file and restart the Workstation. export VMWARE_USE_SHIPPED_GTK="force" The interface will look slightly odd, because older version of GTK is being used, but at least it will work properly. Note: After upgrading a new Linux kernel, it is necessary to compile the VMware modules, this requires to temporarily comment the export line in /etc/vmware/bootstrap.

    Read the article

  • Create Virtual Image of Laptop before Formatting

    - by Simon Mark Smith
    I have a 3 year old laptop running Windows XP that I used for business. Although I have not used the laptop in over a year, I now want to re-commission it with Windows 7 and a fresh install. Before I do the fresh install I want to create a Virtual Image of the laptop that I can keep and potentially run on my desktop machine should I ever need to access any of the old files/projects that it contains currently. I know that most people will say just copy the files over to your desktop, but my concern is the configuration of the laptop. I used to use it for development and it has older versions of Visual Studio, SQL Server, Active X controls etc, etc than I currently use so I really want to preserve the environment not just the files. So really I am asking what is the best tool-set/method to achieve this? I understand there are free VM tools available but I have never done this before and would appreciate any help.

    Read the article

  • Wireless drops on HP ENVY dv6 with RT3290 wireless, worked without problem prior to upgrading to Ubuntu 13.10, can it be fixed?

    - by Tim
    I have a HP ENVY dv6 Notebook PC with an AMD A10 quad core and RT3290 wireless. Since I upgraded from Ubuntu 13.04 to 13.10, the wireless connects, but then drops after a few minutes or longer, whether or not I am running openconnect to get through a VPN. If I attempt to run a remote X client (e.g. remote xterm) it drops. If I don't run an X client, it disconnects after a while, requiring a reload of the driver and reconnect. Wireless info... sudo lshw -c network *-network description: Wireless interface product: RT3290 Wireless 802.11n 1T/1R PCIe vendor: Ralink corp. physical id: 0 bus info: pci@0000:02:00.0 logical name: wlan0 version: 00 serial: 68:94:23:a7:09:cb width: 32 bits clock: 33MHz capabilities: pm msi pciexpress bus_master cap_list ethernet physical wireless configuration: broadcast=yes driver=rt2800pci driverversion=3.11.0-12-generic firmware=0.37 ip=192.168.1.115 latency=0 link=yes multicast=yes wireless=IEEE 802.11bgn resources: irq:55 memory:f0210000-f021ffff I have successfully built and installed the MediaTek driver with no luck on connecting, then the system hangs on reboot and I have to recover/undo the changes to boot successfully.

    Read the article

  • Adding FTP Users via Simple Control Panel

    - by Aristotle
    I just got setup with CentOS yesterday through GoDaddy, and today I'm trying to get into the server and start setting up some of my projects. I'm able to get in through PuTTY just fine, but I'm not able to connect through FTP with the same (root) account. I'm using Simple Control Panel, and have ensured that "Enable Server" is checked beneathSystem Configuration > FTP. Further, I've checked, and double checked that my root password is correct when providing the FTP details. Is there some other common setting I'm missing here that will prevent me from getting connected to the server to begin transfering files?

    Read the article

  • How to run Firefox jailed without serious performance loss?

    - by Vi
    My Firefox configuration is tricky: Firefox runs at separate restricted user account which cannot connect to main X server. Firefox uses Xvfb (virtual "headless" X server) as X server. x11vnc is running on that Xvfb. On the main X server there is vncviewer running that connect to this x11vnc On powerful laptop (Acer Extensa 5220) it seems to work more or less well, but on "Acer Aspire One" netbook it is slowish (on a background that firefox is loaded with lots of extensions). How to optimise this scheme? Requirements: Browser cannot connect to main X server. Browser should be in chroot jail (no "suid" scripts, readonly for many things) Browser should have a lot of features (like in AutoPager, NoScript, WoT, AdBlockPlus)

    Read the article

  • Recovering data from /

    - by Abhijit Gavas
    I accidentally installed Ubuntu to one of my data drives from Windows. The drive was a NTFS drive and contained about 80 GB of important data. The size of the drive is 110 GB. Its new file system is ext4. In an attempt to recover the data, I downloaded foremost and tried the following commands: foremost -i / -o /media/281C8DB01C8D7998/Recovery/ -T -v foremost -i /dev/sda7 -o /media/281C8DB01C8D7998/Recovery/ -T -v (sda7 is the drive in question.) It appears that with either command, foremost gets stuck reading some file. Here is the console output: abhi@abi-PC:/dev$ foremost -i /dev/sda7 -o /media/281C8DB01C8D7998/Recovery/ -T -v Foremost version 1.5.7 by Jesse Kornblum, Kris Kendall, and Nick Mikus Audit File Foremost started at Fri Sep 28 20:58:00 2012 Invocation: foremost -i /dev/sda7 -o /media/281C8DB01C8D7998/Recovery/ -T -v Output directory: /media/281C8DB01C8D7998/Recovery_Fri_Sep_28_20_58_00_2012 Configuration file: /etc/foremost.conf Processing: stdin |------------------------------------------------------------------ File: stdin Start: Fri Sep 28 20:58:00 2012 Length: Unknown Num Name (bs=512) Size File Offset Comment Killed As you can see I have to kill it from system monitor. This approach does not seem to be working. What else could I try to recover the files? Please help. The files are very important and I will be devastated if I cannot recover them.

    Read the article

  • install grub on disk image

    - by Dima
    I have disk image with 2 partitions: Partition 1 has cramfs file system (read only). This partition contains all system files of the OS Partition 2 has ext3 file system. This partition has only configuration files that may be changed. How can I install GRUB1 boot loader on MBR. I tried to copy first 446 bytes of my hard disk and copy GRUB files to the /boot directory on the 1st (cramfs) partition. I cannot use grub-install because I have disk image and not disk itself. Any ideas?

    Read the article

  • Firewalling a Cisco ASA Split tunnel

    - by dunxd
    I have a Cisco ASA 5510 at head office, and Cisco ASA 5505 in remote offices. The remote offices are connected over a split tunnelled VPN - the ASA 5505s use "Easy VPN" Client type VPN in Network Extension Mode (NEM). I'd like to set firewall rules for the non-tunnelled traffic only. Traffic over the VPN to head office should not have any firewall rules applied. I might want to apply different firewall rules to different remote offices. All the documentation I have been able to find assumes the Client VPN is a software endpoint, and all the configuration is done at the 5510. When using a Cisco 5505 as the VPN client, is it possible to configure any firewalling at the Client end, or does it all have to come from the 5510? Are there any other issues to look out for when split-tunnelling a VPN by this method?

    Read the article

  • Windows 7 stopped connecting to samba share

    - by Jan
    I have a problem with Windows 7 Home connecting to samba on linux server. It's worked fine for a year and without any changes in configuration or updates it suddenly stopped working. The strangest thing for me is that it stopped working for no reason that I could think of. After a few minutes after trying to connect I get: \\ip is not accessible. You might not have permission to use this network resource. Contact the administrator of this server to find out if you have access permissions. Other computers in that network (XP machines) connect to samba without any problem. I've tried adding LmCompatibilityLevel as described here: http://www.tomshardware.com/forum/75-63-windows-samba-issue , but it didn't help.

    Read the article

  • Everytime I click on a link from within Mac, parallels boots up my Windows VM and opens the link there

    - by Christoph
    Everytime I click on a link from within Mac, parallels boots up my Windows VM and opens the link there. That's super annoying and I haven't found a configuration to change it. In the settings of my Windows VM there is an option where to launch the webbrowser. However that seems to only affect links that I open from within my Windows session. UPDATE I just stumbled across this answer: How to disassociate a file type with an application in Mac OS So I disabled "share windows apps with mac" completely. Also this seems to be overly protective it at least does the trick. I think it also deactivates that running windows apps show up in the dock but I don't care that much about that anyway.

    Read the article

  • Apache reverse proxy, redirect requests based on IP addresses

    - by Mr Aleph
    I have a Linux box with 2 NICs. I installed and configured Apache 2 for reverse proxy. Each NIC has its own IP address and I was wondering if there is any way to redirect the requests via Apache based on the IP address that was used to get to the box. For example: eth0 has IP 100.100.100.100, eth1 has 200.200.200.200 If I browse to http://100.100.100.100/AppName/App I want it to redirect to 1.1.1.1 and if I browse to http://200.200.200.200/AppName/App I want it to go to 2.2.2.2 Right now the configuration for Apache is set as follow ProxyPass /AppName/App http://1.1.1.1/AppName/App ProxyPassReverse /AppName/App http://1.1.1.1/AppName/App So anything going to /AppName/App will be redirected to 1.1.1.1 I was reading something about ProxyHTMLURLMap but I don't know whether this is something that might help. Any idea how to do this? Thanks!

    Read the article

  • How to netboot ubuntu running iniside VirtualBox on Mac Air

    - by murungu
    Having configured a virtual machine for Ubuntu on VirtualBox on my mac air I need to install Ubuntu OS itself. I have selected the hardrive as the primary boot device and the network as the secondary boot device, so I am not prompted to install an Ubuntu disk at boot time. It attempts to netboot but is unable to locate Ubuntu and cannot find anywhere in the configuration where I can explicitly specify where to find and Ubuntu image, so assume it reverts to some default location and fails. Has anybody out there ever successfully installed ubuntu on virtual box on their Mac Air? What steos did you take to get it right?

    Read the article

  • Why was my site rejected for Google Adsense?

    - by hyuun jjang
    I have a 3 year old blog and its got around 16 articles/tutorials about some programming problems and solutions. It's getting pretty much a lot of view lately so I decided to apply for a google adsense account. When I first applied via blogger, google replied with the following statement: Page Type: In order to participate in Google AdSense, publishers' websites and application information must satisfy the following guidelines: - Your website must be your own top-level domain (www.example.com and not www.example.com/mysite). - You must provide accurate personal information with your application that matches the information on your domain registration. - Your website must contain substantial, original content... So, as I understood it, I decide to buy a domain and point my blogger blog to that new naked domain. and here is the newly bought domain where all the contents of my old blog resides. http://icodeya.com/ I reapplied, hoping that this time, I will make the cut. But then I got this reply Further detail: Unable to review your site: While reviewing http://www.icodeya.com/, we found that your site was down or unavailable. We suggest you check whether there was a typo in the URL submitted. When your site is operational, you can resubmit your application with the correct site by following the directions below. I'm a bit disappointed. Maybe I did something wrong with DNS configuration or something. But you can clearly see that my site is fully functional. I heard that google sends robots to crawl on to the site etc. It's just sad because I invested on a domain name, and now I can't even find ways to earn from it. Any tips?

    Read the article

  • Running Ubuntu with multiple monitors having a panel each

    - by Andreas Grech
    I have just installed Ubuntu and I'm trying to set up a correct configuration for having multiple monitors. I want the two monitors to have their own workspaces, yet the ability to drag a window from one monitor to the next. I also want a separate panel (taskbar) for each monitor. I have played around with nvidia-settings and tried the following different configs: Having both monitors `Use separate X screen" A panel for each monitor but I couldn't drag windows from one monitor to the next Using TwinView The ability to drag windows from one monitor to the next but only a single panel for the primary display Now I want to have the above configurations combined. Basically, as I stated beforehand, having a separate panel for each monitor and also the ability to drag windows across monitors. Can anyone provide me with suggestions please?

    Read the article

  • Win7 OpenSSH config: no address associated with name

    - by Jonah
    I am using OpenSSH on win7. My home dir is C:\Users\JG, and inside that dir I have the file C:\Users\JG\.ssh\config, with these contents: Host <redacted server ip here> HostName digitalocean_git User git IdentityFile ~/.ssh/digitalocean_moocho/id_rsa The id file pointed to by the "IdentityFile" entry works, as I use it just fine via putty, but for this problem I am trying to get command line OpenSSH working. The crux of the problem is explained by this output: >ssh -v digitalocean_git OpenSSH_4.6p1, OpenSSL 0.9.8e 23 Feb 2007 debug1: Reading configuration data /c/Users/JG/.ssh/config ssh: digitalocean_git: no address associated with name Why is no address associated with the name? How can I make this work?

    Read the article

  • Installing Collective Access

    - by Michele
    I am VERY new to installing any type of server program and to running any opensource type software in general. I am running Windows Server 2008R2. I want to install Collective Access to run locally only on my Intranet at home. So my host is localhost I sucessfully installed PHP and MYSQL. I installed CA in this directory C:/inetpub/wwwroot/collectiveaccess. 1st. I do not want to send mail through collective access. Will it install without all the email information? Can I comment those requirements out in the global config and setup.php file? 2nd I am getting the error. Configuration file is missing for hostname 'localhost' this is what I have in the set up file: define("CA_WEB_ROOT_DIR", "c:inetpub/wwwroot"); define("CA_URL_ROOT", "/collectiveaccess"); define("CA_SITE_HOSTNAME", "localhost"); define("CA_DB_HOST", 'localhost');

    Read the article

  • Timeout Error in SQL Server Database Mail Feature

    - by RedLEON
    I configured database mail profile as gmail smtp server, SSL and port 465. I didn't restart server. And this is first time to use Database Mail feature on that server. When I send a testing mail it give me this error message: The mail could not be sent to the recipients because of the mail server failure. (The operation has timed out) I tried this mail configuration with Thunderbird and I could send messages througt this SMTP. Why is SQL Server giving this eror message? I searched here but didn't find any solution.

    Read the article

  • Should `keepalive_timeout` be removed from Nginx config?

    - by Bryson
    Which is the better configuration/optimization: to explicitly limit the keepalive_timeout or to allow Nginx to kill keepalive connections on its own? I have seen two conflicting recommendations regarding the keepalive_timeout directive for Nginx. They are as follows: # How long to allow each connection to stay idle; longer values are better # for each individual client, particularly for SSL, but means that worker # connections are tied up longer. (Default: 65) keepalive_timeout 20; and # You should remove keepalive_timeout from your formula. # Nginx closes keepalive connections when the # worker_connections limit is reached. The Nginx documentation for keepalive_timeout makes no mention of the automatic killing, and I have only seen this recommendation once, but it intrigues me. This server serves exclusively TLS-secured connections, and all non-encrypted connections are immediately rerouted to the https:// version of the same URL.

    Read the article

  • Battery not recognized on my laptop (and it recognizes my laptop as a desktop)

    - by AZorin
    I have installed Ubuntu (both 10.10 and 11.04 pre-release) on my laptop but my battery is not recognized and it is detected as a desktop system rather than a laptop. I have tried to get the output of cat /proc/acpi/battery/BAT1/state but the directory doesn't exist. I have tried another guide to paste the battery info into this directory but it doesn't allow me to do that and says that the directory doesn't exist, even though I'm trying to make it. I tried it in root Nautilus and even on an install of Lubuntu (with a root file manager) but it still failed to budge. I really don't know what to do as I have tried all the guides on the internet that I could find. Is there any way to change the configuration file(s) that detect the internal hardware of the computer. The /proc directory is a temporary RAM directory afaik. Is there a directory where that data is stored permanently and where the RAM reads if you know what I mean? Thanks in advance. AZorin This issue has been reported as bug #764513.

    Read the article

  • Is there a secure way to add a database troubleshooting page to an application?

    - by Josh Yeager
    My team makes a product (business management software) that our customers install on their own servers. The product uses a SQL database for data storage and app configuration. There have been quite a few cases where something strange happened in the customer's database (caused by bugs in our app and also sometimes admins who mess with the database). To figure out what is wrong with the data, we have to send SQL scripts to the customer and tell them how to run them on the database server. Then, once we know how to fix it, we have to send another script to repair the data. Is there a secure way to add a page in our application that allows an application admin to enter SQL scripts that read and write directly to the database? Our support team could use that to help customers run these scripts, without needing direct access to the SQL server. My big concerns are that someone might abuse this power to get data they shouldn't have and maybe to erase or modify data that they shouldn't be able to modify. I'm not worried about system admins, because they could find another way to do the same thing. But what if someone else got access to the form? Is there any way to do this kind of thing securely?

    Read the article

  • Multiple SSIDs better or worse

    - by swiss196
    I'm just setting up a network in a student house with 10 rooms, 3 floors. Configuration at the moment is as follows: Virgin Media 100mb Cable Virgin SuperHub on ground floor broadcasting on SSID1 Second AP(Edimax) wired on middle floor broadcasting on SSID2 These two networks server all 3 floors fairly well but I was wondering whether it would be better to setup an individual SSID for each floor (both the routers allow me to configure up to 3 ssids on each!). Would this help with speed issues, i.e if someone on the top floor is downloading/streaming etc, this wouldn't affect someone on the middle floor on a different SSID Or, will it have no effect? Thanks, Dave

    Read the article

  • Effects of enabling DVFS in a CPU

    - by Nrew
    I came up with this energy regulating software called Granola while reading at Ghacks.net I downloaded and installed the software but you need to enable the DVFS function in the CPU. I booted up and tried to enable DVFS in the CPU Configuration in my Asus motherboard(Desktop). But in the description it said that enabling the feature might raise some compatibility issues with the power supply. And I'm afraid that I might ruin the CPU if I do this. Any suggestions, any one here who has tried this before? I don't know who the manufacturer of my power supply is. Would you recommend some software that can determine every possible info about the power supply.

    Read the article

  • 14.04 LTS Unity no longer boots after last 94 MB update

    - by Harryg123
    I am running 14.04 LTS, Unity, on an HP Pavillion 15, 4GBRAM, 750 GB hdd, I-5 machine, with AMD 8600M graphics card built in. I have disabled the dash and all Ubuntu spyware. I have been faithfully loading all updates as they appear. This morning it asked for a 94 MB update (bringing kernel to .27, I think. Now, I can boot, get to login screen, but it freezes after that. Keyboard doesn't work at that point, but mouse does. I booted into recovery mode, tried to run in generic graphics mode, -- system again froze. I also pressed [esc] during boot, but saw nothing strange; then text disappeared and was replaced by login screen. I am not a hobbyist; this is a production machine and I have a lot of work to do today. Having a standard software update render my machine completely useless... sigh. Perhaps the simplest thing to do would be to revert to the previous configuration. How do I do that? I can boot into recovery mode. But I have no idea how to proceed. TIA for all help. -Harry

    Read the article

  • How can I tell GoogleBot that a subdirectory is now a subdomain?

    - by cwd
    I had about a million pages of a catalog indexed under a subdirectory, and now that's moved to a subdomain. GoogleBot is crawling each one of them and getting a 301 redirect to the new location. Even though I have set up the redirect rule in the apache sites-enabled configuration file, (i.e. it's early on when apache does the redirect - PHP is not even getting loaded), even though I have done that, the server isn't handling the load well. GoogleBot is making around 5 requests per second, and on top of my normal traffic that is hiking up the CPU for a few hours at a time. I checked in Webmaster Tools and the corresponding documentation for a way to let Google know that the content had been moved from a subdirectory to a subdomain, but with little luck. Basically the most helpful thing I saw said to just send 301 headers for the new location. How can I tell GoogleBot that a subdirectory is now a subdomain? If that is not an option, how can I more efficiently send 301 redirects out for a particular subdomain? I was thinking perhaps the Nginx server but I'm not sure that I can run both Apache and Nginx side by side on port 80 for different subdomains.

    Read the article

  • The application attempted to perform an operation not allowed by the security policy

    - by user16521
    I ran this command on the server that has the share of code that my local IIS site set to (Via UNC to that share): http://support.microsoft.com/kb/320268 Drive:\WINDOWS\Microsoft.NET\Framework\v2.0.50727\caspol.exe -m -ag 1 -url "file:////\\computername\sharename\*" FullTrust -exclusive on (obviously I replaced Drive with C, and the actual computername and sharename with the one I'm sharing out). But when I run the ASP.NET site, I am still getting this runtime exception: Description: The application attempted to perform an operation not allowed by the security policy. To grant this application the required permission please contact your system administrator or change the application's trust level in the configuration file. Exception Details: System.Security.SecurityException: Request for the permission of type 'System.Web.AspNetHostingPermission, System, Version=2.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089' failed.

    Read the article

< Previous Page | 365 366 367 368 369 370 371 372 373 374 375 376  | Next Page >