Search Results

Search found 18927 results on 758 pages for 'remote desktop'.

Page 508/758 | < Previous Page | 504 505 506 507 508 509 510 511 512 513 514 515  | Next Page >

  • Can I connect an external antenna to a range extender?

    - by ercan
    I live in a kind of dormitory and the next access point is 60 meters away from my room. So I bought a range extender (TP-Link WA730RE) and installed it into my room in the same height as the access point. But my problem is still not solved. The reception is slightly better but my connection still gets broken every two minutes. The antenna that comes with this range extender was 4dB. My question is, can I buy an 8dB external antenna (like this one: http://www.amazon.com/TP-Link-TL-ANT2408C-Desktop-Omni-Directional-Antenna/dp/B0034CQSKW/ref=sr_1_14?ie=UTF8&qid=1332074993&sr=8-14) and replace it with the antenna that comes with the range extender? Or is this external antenna only suitable for the "receiver" end of the connection, i.e. the computer?

    Read the article

  • Persistent network share connection not working with runas

    - by binarycoder
    If I use runas /user:DOMAIN\user cmd.exe (using XP), previously mapped persistent network drives are considered unavailable. net use shows: Status Local Remote Network ------------------------------------------------------------------------------- Unavailable H: \\SERVER\SHARE Microsoft Windows Network dir H: fails with "The system cannot find the path specified.". The connection is easily revived with `NET USE H: \SERVER\SHARE': not asked for a password when I do this. What is going on? Can I make Windows safely revive this drive automatically when it is first accessed.

    Read the article

  • squid configuration change to accept http request on LAN

    - by Ratan Kumar
    installed squid + dansguardian to block adult content on my linux (ubuntu 12.10) . everything worked fine. it has blocked as expected . now the problem is i am also running an apache server for my LAN . ( kind of website ) but when accessing it via 192.168.0.1 , it says squid has blocked the connection , this is the exact error The following error was encountered while trying to retrieve the URL: http: //192.168.0.16/ Connection to 192.168.0.16 failed. The system returned: (113) No route to host The remote host or network may be down. Please try the request again. Your cache administrator is webmaster. before configuring the squid it was working fine . what changes in the squid.conf i have to make . i tried acl Safe_ports 80 allow_all Safe_ports ( i want to know how i can configure it again to listen HTTP request from LAN )

    Read the article

  • How do I get Chrome to not try to open in Windows 8 mode?

    - by Citrus Rain
    It's not that I'm trying to make it more like 7. That's not the problem. My problem is that whenever I click the Google Chrome icon on my taskbar or even the desktop, Windows brings up this screen-spanning message box saying that my resolution is too low to launch run Google Chrome, and to set my computer to a higher resolution. (I can't.) It then continues to open Chrome correctly after I hit close. This also happens when I click a link in an external program like skype or psi.

    Read the article

  • Secondary Monitor won't Display at Full Resolution with Primary Display Disconnected

    - by Laramie
    I have had a bad day. Among today's events, the LCD display on my Dell E6500 notebook failed. It had been working with a secondary monitor connected through a VGA cable at 1600x900 resolution until I removed the LCD from my notebook for repair. Now the max resolution available on the secondary display is 1024x768. If I reconnect the burned out primary display, the secondary monitor is again available at 1600x900. The video card is an onboard Mobil Intel 4. How can I maintain the secondary display at 1600x900 with the primary display disconnected (for repair)? Edit: I just noticed that when Windows boots, it displays at the correct resolution and I can see my desktop in all its glory for a couple seconds then reverts to 1024x768.

    Read the article

  • What is the simplest, open-source webmail frontend available?

    - by josePhoenix
    I am working on a project to create a few extremely stripped down interfaces for common Web/Internet tasks in order to make computers accessible to my visually impaired grandmother. Currently she uses Mac OS X Mail.app, but I had the idea that I could re-skin a webmail interface running on my own server to make it easier for her to use. The ideal webmail interface to use as a starting point would be without frames or AJAX and written in Python, Perl, or PHP5+, though any setup could work as long as the template and stylesheet files were separate from the application itself. This frontend must also connect to a remote IMAP server, since her email account is with her ISP and not on my server. Can anyone recommend a bare-bones, no-nonsense webmail interface that would work for this?

    Read the article

  • Worst SysAdmin Accident

    - by Ward
    In line with the question about Best sysadmin accident, what's the worst accident you've been involved in? Unlike the previous question, I mean "worst" in the sense of most system damage or actual harm to people. I'll start with mine: We have two remote wiring closets that are at the end of a 100-foot corridor which has a metal grate for the floor. After we had Cat6 cable installed, the contractors cleaned up all the debris that dropped through the grating to the concrete 3 feet below. A co-worker and I entered the corridor to check on the progress one day but were distracted and didn't notice that a piece of grating had been moved aside. My buddy stepped into air and his chest slammed into the steel crossbar. He was winded and sore enough to take a couple days off, but luckily the steel beam had rounded edges and the size of the opening was such that he didn't smack his head into it or the floor below. Obviously we learned that areas where the floor is partially removed need to be flagged.

    Read the article

  • Utility to add visual effects to Windows

    - by Soumya92
    I have no idea how to go about looking for this utility, so I though I would ask here. I recall a small utility that would add Compiz like effects on Windows. It was relatively small, and was released a few years ago. Does anyone know what it was called, and where I can get it? EDIT: It does not add the desktop cube or window switchers. It is more like the Compiz panel, where you can check/uncheck the effects that you want. The effects are applied to windows, and is stuff like fade, blur, etc.

    Read the article

  • How can I dual install Ubuntu 10.4 in a Mac Mini with 10.4.11?

    - by Marco Mariani
    I'd like to power-up my aging Mac Mini (1.5GHz Core Solo, 1GB RAM, Tiger 10.4.11) by installing a shiny Ubuntu alongside the current OS. After all, I use Ubuntu for everything save for cleaning my teeth. Since it's my first and only Mac and I have next to no experience with the OS (having used it basically as a media player) I am a little concerned about rEFIt, ELILO, Boot Camp and the fact that it's basically a 4.5 years old unsupported machine and I might get asleep reinstalling everything several times. I've used the live desktop-i386 CD and everything works. I tried with an external USB drive instead of a CD but couldn't make it boot. As for installing Ubuntu, the howtos I've found give several alternatives depending on the model, the OSX version, etc.. but they usually talk about newer machines. Which howto should I follow to repartition, and boot thereafter? Thanks

    Read the article

  • Site to site VPN using RRAS from an untrusted network?

    - by DrZaiusApeLord
    Our remote office will be moving to a new space where internet will be provided. They'll be behind a router doing NAT (I do not have admin rights to this router). They will be sharing a printer with the other people on the LAN, but will need VPN to our network for email and file shares. I was thinking of just having them run the windows VPN client and connecting via PPTP like they do when they are off-site, but I have read that multiple PPTP connections from the same NAT'd address to the same destination doesn't work well or at all. I am thinking some kind of site-to-site VPN is needed so there is just one tunnel. Can I just put in a VPN gateway, set it to connect to our RRAS/PPTP server, and have them use it as their default gateway? Perhaps even use the local default gateway for internet traffic. If so, what VPN gateway/device is recommended for this? Or other solutions? Thanks.

    Read the article

  • mail server administration

    - by kibs
    MY postfix does not show that it is listening to the smtp daemon getting mesaage below: The message WAS NOT relayed Reporting-MTA: dns; mail.mak.ac.ug Received-From-MTA: smtp; mail.mak.ac.ug ([127.0.0.1]) Arrival-Date: Wed, 19 May 2010 12:45:20 +0300 (EAT) Original-Recipient: rfc822;[email protected] Final-Recipient: rfc822;[email protected] Action: failed Status: 5.4.0 Remote-MTA: dns; 127.0.0.1 Diagnostic-Code: smtp; 554 5.4.0 Error: too many hops Last-Attempt-Date: Wed, 19 May 2010 12:45:20 +0300 (EAT) Final-Log-ID: 23434-08/A38QHg8z+0r7 undeliverable mail MTA BLOCKED OUTPUT FROM lsof -i tcp:25 command master 3014 root 12u IPv4 9429 TCP *:smtp (LISTEN) (Postfix as a user is missing )

    Read the article

  • Hiera + Puppet classes

    - by Amadan
    I'm trying to figure out Puppet (3.0) and how it relates to built-in Hiera. So this is what I tried, an extremely simple example (I'll make a more complex hierarchy when I manage to get the simple one working): # /etc/puppet/hiera.yaml :backends: - yaml :hierarchy: - common :yaml: :datadir: /etc/puppet/hieradata # /etc/puppet/hieradata/common.yaml test::param: value # /etc/puppet/modules/test/manifests/init.pp class test ($param) { notice($param) } # /etc/puppet/manifests/site.pp include test If I directly apply it, it's fine: $ puppet apply /etc/puppet/manifests/site.pp Scope(Class[Test]): value If I go through puppet master, it's not fine: $ puppet agent --test Could not retrieve catalog from remote server: Error 400 on SERVER: Must pass param to Class[Test] at /etc/puppet/manifests/site.pp:1 on node <nodename> What am I missing? EDIT: I just left the office but a thought struck me: I should probably restart puppet master so it can see the new hiera.conf. I'll try that on Monday; in the meantime, if anyone figures out some not-it problem, I'd appreciate it :)

    Read the article

  • System Center Configuration Manager 2007 - Debugging Client Installs

    - by Dayton Brown
    Hi All: Having an issue installing the CCMsetup client on desktops. The CCMSetup makes it to the PC, files are there, it gets added to the services for automatic start, it starts, but quits almost instantly. Logs on the desktop show a entry like this. <![LOG[Failed to successfully complete HTTP request. (StatusCode at WinHttpQueryHeaders: 404)]LOG]!><time="14:28:51.183+240" date="06-11-2009" component="ccmsetup" context="" type="3" thread="2388" file="ccmsetup.cpp:5808"> What am I missing? EDIT: Firewall is off on both client and server.

    Read the article

  • continuous hard disk access - slowing down my machine

    - by suresh
    I find from the hard disk access LED on the front of my machine that the hard disk is being accessed more often and probably because of that, my machine is quite slow. The machine becomes unresponsive even when the load as seen from w command is around 1 or so. My desktop is optiplex 360 dell machine running Ubuntu 10.04. My questions are: How to quantify hard disk access and how to see if it is more than "normal" ? If it is more than normal, what are my solutions ? thanks suresh

    Read the article

  • wmpnetwork.exe service clogs CPU usage

    - by Brenton Taylor
    We have remote locations each with 2 ASUS media extenders that stream from a computer with a shared Media Player library. Lately, several of these locations experience the "wmpnetwork.exe" service throttling the CPU to 100% usage. Killing the service only results in it starting back up, and so far the only temporary solution is to uninstall Media Player. A lot of these computers are also about 3-5 years old. Could it just be a case of outdated hardware not being able to do everything we ask them to do? Edit: all running Windows XP and Windows Media Player 11

    Read the article

  • Security issues of running PHP scripts as the owner of the PHP file with suexec

    - by thomasrutter
    I'm using suexec to ensure that PHP scripts (and other CGI/FastCGI apps) are run as the account holder associated with the relevant virtual host. This allows for securing each users' scripts from reading/writing by other users. However, it occurs to me that this opens up a different security hole. Previously, the web server ran as an unprivileged user, with read-only access to user's files (unless the user changed the file permissions for some reason). Now, the web server can also write to user's files. So while I've prevented different users taking advantage of each other's scripts, I've made it so that in the event that some application has a remote code injection vulnerability, it now has not only read access but also write access to all that user's scripts and website. How can I deal with this? One idea I've had is to create a second user account for each user account in the system, so that each user has their own user account, and all their scripts are run under another user account. But that seems cumbersome.

    Read the article

  • Running PHP scripts as the owner of the PHP file: security issues

    - by thomasrutter
    I'm using suexec to ensure that PHP scripts (and other CGI/FastCGI apps) are run as the account holder associated with the relevant virtual host. This allows for securing each users' scripts from reading/writing by other users. However, it occurs to me that this opens up a different security hole. Previously, the web server ran as an unprivileged user, with read-only access to user's files (unless the user changed the file permissions for some reason). Now, the web user can also write to user's files. So while I've prevented different users taking advantage of each other's scripts, I've made it so that in the event that some application has a remote code injection vulnerability, it now has not only read access but also write access to all that user's scripts and website. How can I deal with this? One idea I've had is to create a second user account for each user account in the system, so that each user has their own user account, and all their scripts are run under another user account. But that seems cumbersome.

    Read the article

  • Bare bones backup / restore for single Win 2003 server

    - by s.mihai
    I have a single server, Win 2003 Server and would like to setup a system to be able to perform a bare bones restore if needed. (just plug in a cd or smth and get everything back). Ideally the backup could be performed while the is powered so that i don't have to support downtime during this, and in order to restore i would reboot and use some sort of liveCD Any ideas on this, software and all... ? (backup will be done to a remote FTP server with plenty of bandwidth).

    Read the article

  • Installing a Wl-138g wireless NIC on ubuntu server

    - by Narcolapser
    So I have this little server that I've been working on as a hobby. But due to changes in living arrangements I have move and it won't have access to a wired lan. So I need to install a wireless adapter i had sitting around. But most tutorials I have run across are either out dated, or about ubuntu desktop and talk about things using a GUI. So my question is firstly is there a utility for identifying and downloading drivers for new hardware. that would save my having this problem in the future. If not, how do i go about installing said wireless NIC? ~N

    Read the article

  • setting up delegate or smtp forwarding

    - by cotiso
    for work we have a remote dedicated server to run our webservice that also runs our email services, at home(comcast residential internet) i cannot send mail using the dedicated server's SMTP, comcast spits back a error saying i can only use their SMTP server for sending mail at work(comcast business internet) we can use our dedicated server for sending mail with no problem so i set up a box at work to forward smtp traffic, i'm new to all this networking stuff by the way i used delegate to forward smtp traffic, can someone point me in the right direction on how to use this program(delegate) to fix our issue the delegate command i used to test is : delegated -P25 SERVER="smtp://dedicated.server.com:25" PERMIT=":::" -v i also opened up port 25 on the router so it points to my boxes ip are there any other ways to fool comcast into thinking im using my works ip to send mail, my coworkers and i are unable to send mail from home for some time now thanks

    Read the article

  • How can I share a video file during a webinar?

    - by Brien Malone
    Here is the scenario: I have a number of remote employees around the globe. I want to have a video chatting session. No problem there. Halfway through, I want to shut off all camera video feeds and simulcast (synchronous) a training video to my team. How do I do this? We have tried office communicator, but the frame rate was awful and no audio. Adobe Connect had similar trouble. In both cases we were limited by the main office's small internet pipe, but it is clear that video delivered by shared desktop is not a good solution.

    Read the article

  • Why screen freezes in Windows 7 and Vista?

    - by Hasan Gürsoy
    On Windows 7 and Vista I use auto-hide task bar and turn off graphic card after 1-2 minutes. When I want to use my desktop again screen freezes sometimes for 10 seconds. What I do does not effect anything immediately but cursor moves. For example if I click Firefox bookmark it opens when screen unfreezes. This happens at any of my computers and about a month after new installation. I don't think that problem is related with the drivers.

    Read the article

  • Increase backup speed, Backup Exec 2010 - QNAP TS419U+ NAS

    - by user99912
    We have a QNAP NAS and the network shares are being backed up by Backup Exec 2010 over SMB. We can't install the remote agent on the NAS as it has an ARM processor and, as far as I am aware, there is no compatible agent. Do you have any suggestions on any faster method of backing up these shares as opposed to the current scenario? Currently the network bandwidth is not the issue, it seems that this access method is just not able to go any quicker. We've also added the NAS shares to the start of the selection list, but we're still running into 18 hours total backup time (total amount of data on the NAS is roughly 650GB). Any comments and/or suggestions welcome. EDIT: Data is being pulled from the NAS by Backup Exec to a LTO4 tape drive

    Read the article

  • Puppet master/agent basic setup

    - by lewap
    I'm trying to setup a basic puppet agent/master use-case with an agent server and a master. I've setup two servers with puppet and puppet master respectively. After the following setup of both servers: puppet master --no-daemonize --verbose puppet agent --test puppet cert --list to get the list, puppet cert --sign to sign it. puppet agent --test I get the message: err: Could not retrieve catalog from remote server: hostname was not match with the server certificate warning: Not using cache on failed catalog err: Could not retrieve catalog; skipping run err: Could not send report: hostname was not match with the server certificate What do I need to do in order to get the agent/master to be able to talk to each other?

    Read the article

  • Are periodic full backups really necessary on an incremental backup setup?

    - by user2229980
    I intend to use an old computer I have as a remote backup server for myself and a few other people. We are all geographically separated, and the plan is to do incremental daily backups using rsync and ssh. My original idea was to make one initial full backup then never again have to deal with the overhead of doing it, and from that moment on only copy the files changed since the last backup. I've been told that this could be bad, but I fail to understand why. Since each snapshot is comprised of hard links to the unchanged files plus the original changed ones, isn't it going to be identical to a new full backup? Why would I want to make another full backup?

    Read the article

< Previous Page | 504 505 506 507 508 509 510 511 512 513 514 515  | Next Page >