Search Results

Search found 25959 results on 1039 pages for 'remote support'.

Page 113/1039 | < Previous Page | 109 110 111 112 113 114 115 116 117 118 119 120  | Next Page >

  • DOS application to allow remote management of files over serial link

    - by tomlogic
    Harken back to the days of DOS. I have an embedded DOS handheld device, and I'm looking for a tool to manage the files stored on it. I picture an application I can launch on the device that opens COM1 up for commands to get a directory listing, send/receive files via x/y/zmodem, move/delete files, and create/move/delete directories. A Windows application can then download a recursive file listing and then manage those files (for example, synchronizing with a local directory). Keep in mind that this is DOS -- 8.3 filenames, 640K of RAM and a 19200bps serial link (yuk!). I'd prefer something with source in case we need to add additional features (for example, the ability to get a checksum of a file for change detection). Now that I've written this description, I realize I'm asking for something like LapLink or pcAnywhere. Norton no longer sells DOS versions of pcAnywhere and LapLink V for DOS seems pricy at $50. Are you aware of any similar apps from those good old days?

    Read the article

  • Lost support for Web Access on Verizon BlackBerry World Edition

    - by Jimsmithkka
    Hello all, I believe that some silliness has occurred with my blackberry after a OS upgrade. I have 2010 Blackberry world edition phone, purchased off a friend who went iPhone, that at first worked with web on the Verizon network. When i connected it to my PC to transfer contacts, it prompted for an OS upgrade, which I performed. Post-Upgrade I have found that i can no longer access any of the web services: eg. AppWorld, Email, Twitter, Browser. And they all state that i need to upgrade my account to gain access. I had a Storm previous to this that worked fine, and at the VZ store they told me this device is no longer supported (new in 2010 though), and they got me a free "upgrade" to the Blackberry flip. What i could use help with is finding a source stating it is discontinued or a guide that will help me re-enable the web features. I can provide further info later if needed (currently at work with the flip, the WorldEdition is at home).

    Read the article

  • Remote network traffic not passing through VPN

    - by John Virgolino
    We have the following topology: LAN A LAN B LAN C 10.14.0.0/16 <-VPN-> 10.18.0.0/16 --- SONICWALL <-VPN-> M0N0WALL --- 10.32.0.0/16 Traffic between LAN A and LAN B works perfectly. Traffic between LAN C and LAN B works perfectly. Traffic between LAN A and LAN C, not so much. LAN A's gateway has a route to LAN C that points to the Sonicwall. The Sonicwall has a route to LAN A pointing to the VPN gateway connecting LAN B to LAN A. Tracing packets on the Sonicwall shows the LAN C destined traffic to arrive on the Sonicwall, but it does not forward the traffic, it dies there. Traffic from LAN B gets forwarded. Tracing packets on the Sonicwall while sending traffic from LAN C destined for LAN A shows nothing. This tells me that the M0N0WALL is not forwarding traffic for the 10.14.0.0 network and the Sonicwall is not forwarding from 10.14.0.0. The SA on the Sonicwall terminates on the WAN ZONE and is defined to use an address group that incorporates both the 10.14.0.0 and 10.18.0.0 networks. The M0N0WALL is configured for the 10.18.0.0 network and I have tried with both a static route to 10.14.0.0 and without on the M0N0WALL. I tried manually adding the 10.14.0.0 network to the SA on the M0N0WALL, but that really aggravated it and the SA never came up, so I reverted. I have checked all the firewall rules to make sure nothing is blocked. All of the Sonicwall auto-added rules look right. Specs: Sonicwall TZ200, Enhanced OS M0N0WALL v1.32 I'm at a loss at this point. Any help would be appreciated.

    Read the article

  • Remote I/O costs with a Content Delivery Network

    - by x711Li
    As far as I know, the time complexity of scanning a directory and the amount of files in said directory are correlated due to I/O costs. Would the administrative costs of placing the files in a hashed directory tree for uploading/downloading files through a CDN API be worth it for the added efficiency? For instance, given a filename foo.mp3, the MD5 hash for this is 10ebb1120767e9de166e0f5905077cb1. Thus, storing foo.mp3 in ./10/eb/foo.mp3 would allow for less files per directory (assuming MD5 generates patterns with in Base36, this allows for 36^2 root directories with 36^2 subdirectories each and little chance of hash collision) Considering the directories themselves are not loaded, would the I/O costs of directory scanning still exist with direct uploading/downloading?

    Read the article

  • Remote connection to Mysql not working

    - by Fillipe Silva
    We have an application running with CodeIgniter and MySQL in cestacerta.com. Ther are two versions: One that is in production and it works perfectly, and one that is our Development version. This runs locally on my machine and the other developers. The development version needs to remotely connect to the database, and it rather suddenly stopped working. I've given permission to access any IP settings on the server. I can access the database through the MySQL Gui Tools. I have tested access to several different codes, including a newly downloaded version of CodeIgniter and always got the same error: "A Database Error Occurred Unable to connect to your database server using the provided settings. Filename: C: \ xampp \ htdocs \ cestacerta \ system \ database \ DB_driver.php Line Number: 119 " What troubleshooting steps can I take determine if the error is on our workstations (which all have the same error) or on our server.

    Read the article

  • Extend Linux Desktop to another X Windows Display

    - by unknown (google)
    Hello, I am a long time Linux user of the Xinerama and other technologies for extending a desktop to multiple monitors. However when I travel with my laptop I miss the multi-monitor support I enjoy at home. Recently I acquired a second laptop for a low price. Both laptops are running Fedora (versions 10 and 11 respectively). I use Gnome as my primary desktop environment. I know about synergy. I use synergy all the time to control the screen of other Windows / Linux systems I use. I would like to know, can I sit both my primary and secondary laptops together and achieve a Xinerama-like extended desktop environment? Ideally I would like to start a GNOME session on my primary laptop. And then start a X-Windows Desktop on my secondary laptop and extend my primary laptop's desktop onto it. I would like to be able to move Windows from the primary desktop to the secondary laptop desktop. Would I need to use synergy to do this with some other bit of X-Windows technology? Or is there X-Windows technology that will do all this for me? I am familiar with X Windows ability to display applications remotely. I am also familiar with Nomachine's NoX.

    Read the article

  • Does Windows Home Server support folder redirection?

    - by Shawn Miller
    Windows provides the ability to redirect specific user folders to server locations, using a group policy extension called Folder Redirection. The Wikipedia article and Microsoft's marketing page for Windows Home Server suggests a tons of great features, but I don't see anything about specifically about Folder Redirection. I currently run a domain controller in the home so that I can push the folder redirection group policy down to all PCs. Windows Home Server looks like a better fit, but I'd hate to give up on the ease of folder redirection to automatically save all documents, music, pictures to the file server. Any thoughts?

    Read the article

  • Logging commands executed by remote shell scripts

    - by user145836
    I've noticed that when running a script that connects to a number of our servers (to essentially run batch commands) that the commands aren't logged in the user's .sh_history or .bash_history files. Is there a place where this is logged (assuming the script itself isn't doing the logging and I'm not tee'ing the output anywhere)? I'm talking specifically about AIX, but I would assume this question applies to all the *nix flavors. Thanks!

    Read the article

  • Windows 7 x64 support for Intel GMA 3650 (or GMA 3600)

    - by Loom
    I recently purchased an Intel D2700 MUD motherboard and I cannot find drivers for the Win7 x64 integrated graphics (Intel GMA 3650 aka PowerVR sgs545). The accompanying CD contains Win7 x32 version only. When I run it I got an error: This computer does not meet the minimum requirements for installing the software. I tried to use online utility Intel Driver Update Utility Graphics. I used Chrome, Firefox, Internet Explorer without success. First, UAC prompt appear, and then endlessly spinning progress-bar with text "Analyzing computer...". The text in UAC prompt is: Program file name: System Requirements Lab Verified publisher: Husdawg, LLC I downloaded this utility (intel_srldetect_4.5.5.0) and started it from my hard disk. I got an error: A network error occured while attempting to read from the file: C:\Users\Loom\Downloads\SystemRequirementsLab_intel_4.5.5.0.msi Standard VGA driver works for this video card but without hardware acceleration: Hardware acceleration is either disabled or not supported by your video card driver, which could slow game performance. Make sure you have the latest video card driver installed and that hardware acceleration is turned on. Where I can get appropriate driver?

    Read the article

  • Remote offscreen rendering

    - by redmoskito
    My research lab recently added a server that has a beefy NVIDIA graphics card, which we would like to use to do scientific computations. Since it isn't a workstation, we'll have to run our jobs remotely, over an ssh connection. Most of our applications require doing opengl rendering to an offscreen buffer, then doing image analysis on the result in CUDA. My initial investigation suggests that X11 forwarding is a bad idea, because opengl rendering will occur on the client machine (or rather the X11 server--what a confusing naming convention!) and will suffer network bottlenecks when sending our massive textures. We will never need to display the output, so it seems like X11 forwarding shouldn't be necessary, but Opengl needs the $DISPLAY to be set to something valid or our applications won't run. I'm sure render farms exist that do this, but how is it accomplished? I think this is probably a simple X11 configuration issue, but I'm too unfamiliar with it to know where to start. We're running Ubuntu server 10.04, with no gdm, gnome, etc installed. However, xserver-xorg package is installed.

    Read the article

  • Does my OS will support these tools ?

    - by Zerotoinfinite
    I am currently using Windows Server 2003 and I want to install Windows XP [SP1 or SP2 or SP3] or may be VISTA. I have may application which I can run on Windows Server 2003 and I am curious to know If I could run the same on XP or Vista. Please help me deciding whether I change my OS or not, here is the List of software and app I want to work on: Visual Studio 2008/2010 SQL Server 2008 ASP.NET MVC and Entity Framework WPF application P.S. : I can create all my application [except WPF] with Windows Server 2003. I have a idea that I can install VS 2008 on XP but not exactly sure about MVC framework and other latest technology stuff.

    Read the article

  • Modem with support for wired 802.1x

    - by menko
    I'm looking at allowing internet access for only a few scattered users on a school network. One method I'm considering is using 802.1x authentication for the modem. Of course there are simpler options like only allowing specific IP or mac addresses, but policy requires something more secure than this. So, are there any modems with 802.1x wired authentication? All my searching on this yields info about wireless.

    Read the article

  • Remote SCCM deployment of Operating Systems

    - by Decad
    I am currently using sccm 2007 for our software deployment and PXE. During this summer I have been tasked with upgrading 2000+ machines from Windows XP to Windows 7. My plan is to use sccm to advertise the Windows 7 task sequence to the machines. However my question is, what is the best way to automate the deployment? Can I make SCCM turn a machine on and make it run an advertised task sequence without having to be in the same room as the machines?

    Read the article

  • Architectural advice - web camera remote access

    - by Alan Hollis
    I'm looking for architectural advice. I have a client who I've built a website for which essentially allows users to view their web cameras remotely. The current flow of data is as follows: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Ftp connection is enabled for the cameras ftp user. Web camera opens ftp connection to server. Web camera begins taking photos. Web camera sends photo to ftp server. On image url request: Server reads latest image on hard drive uploaded via ftp for camera. Server deleted any older images from the server. This is working okay at the moment for a small amount of users/cameras ( about 10 users and around the same amount of cameras), but we're starting to worrying about the scalability of this approach. My original plan was instead of having the files read from the server, the web server would open up an ftp connection to the web server and read the latest images directly from there meaning we should have been able to scale horizontally fairly easily. But ftp connection establishment times were too slow ( mainly due to the fact that PHP out of the ox is unable to persist ftp connections ) and so we abandoned this approach and went straight for reading from the hard drive. The firmware provider for the cameras state they're able to build a http client which instead of using ftp to upload the image could post the image to a web server. This seems plausible enough to me, but I'm looking for some architectural advice. My current thought is a simple Nginx/PHP/Redis stack. Web camera issues post requests of latest image to Nginx/PHP and the latest image for that camera is stored in Redis. The clients can then pull the latest image from Redis which should be extremely quick as the images will always be stored in memory. The data flow would then become: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Camera is sent an http request to start posting images to a provided url Web camera begins taking photos. Web camera sends post requests to server as fast as it can On image url request: Server reads latest image from redis Server tells redis to delete later image My questions are: Are there any greater overheads of transferring images via HTTP instead of FTP? Is there a simple way to calculate how many potential cameras we could have streaming at once? Is there any way to prevent potentially DOS'ing our own servers due to web camera requests? Is Redis a good solution to this problem? Should I abandon PHP/Ngix combination and go for something else? Is this proposed solution actually any good? Will adding HTTPs to the mix cause posting the image to become too slow? Thanks in advance Alan

    Read the article

  • Any consumer routers with Outgoing VPN support?

    - by Brian Lacy
    When I'm working at home, I need to be able to connect to three different outgoing VPNs, two of which happen to use the same internal IP addressing schemes (192.168.0.*). I also need a static address for my VirtualBox VM so I can connect to my testing web server. Are there any routers which will allow me to connect to multiple outgoing VPNs and assign different internal IP addresses through NAT? Is such a thing even possible, or are there alternate solutions available? Thanks!

    Read the article

  • Web-based IMAP client with support for multiple mailboxes

    - by Nils
    I would like to switch from desktop-based e-mail software (Thunderbird) to a web-based solution that I run on my own web server. I have already tried out Roundcube and while it does work reasonably well so far there is one great feature from Thunderbird that seems to be missing - it doesn't allow me to have a unified mailbox for multiple IMAP accounts. Can anybody recommend a web-based IMAP client that has this feature?

    Read the article

  • How to change the download save directory in ktorrent from a remote host

    - by Garethj94
    So I have ktorrent running on a server and I don't have the ability to see the X11 display for the client, and I need to change my download directory since I like to keep everything very organized. So I need a way to change the download directory for the program as a whole, not like decide where to put each and every torrent that I download. This can't be done through the webui preferences so I'm guessing that I'll have to do it through ssh somehow but from what I've read there really isn't a command to use. Also the way that I have the application run is on startup my server runs the command ktorrent, so it will already be running when I want to change the download location so I assume that I will have to restart the program as well. If anyone knows how to do this it would be much appreciated, and I can't think that I'm the only person to want this feature.

    Read the article

  • Remote access and local access same hostname

    - by cpf
    Hi serverfault, I have a server in a clients network, seperated from theirs with a router/firewall, the intention is to have this server available through one hostname (example.com) My idea is to have (at least) a DNS server in the outside, to have outside (out of the clients' network) access the internal server. The problem would at that point be the internal client (PC A) My question: What would I have to do to make something like this work? Is it even possible or already done? The goal is to not have to change anything on either PC A or PC B, while both should access the same "internal server" while surfing to "example.com" Perhaps adding logic to the DNS server would work (Detect the external IP of internal client [PC A] is the same as the IP for example.com - Give the local IP as reply?) Anyhow: Thanks for helping me think on this!

    Read the article

  • Unauthorized computer use via keyboard or remote access?

    - by brydaverambo
    I’m suspecting my computer is being used when I’m not at home. This is happening either physically or remotely. My wireless switch is off. Is there any way possible to detect and/or monitor activity without purchasing software? My settings are being changed as well as passwords (Bios PW was changed and I cannot access Bios settings). I connect via the network cable. Is it possible for someone (in range) to connect to my laptop even if the wireless switch is off? This is a Dell Inspiron 1720 with the WLAN 1395 card. Here’s the kicker. When I try to download freeware for monitoring activity, I am not allowed to do this! ????

    Read the article

  • Remote users on domain share

    - by user19039
    I am trying to share a domain share so only a specific machine with a specific user can connect. I have tried all the normal ways of actually adding the computer to the domain and then using its unc to map a user but to no avail. Is this possible? How do you do this?

    Read the article

  • Citrix and Remote Desktop session showing inconsistency after some time - Windows 2003 Server

    - by Nishant
    We have a lab server RDE becomes blank screen after sometime though it connects . When we login , it just shows the blank desktop with nothing visible , just the wallpaper . Citrix takes long time to connect and if it does ( it takes long time "checking credential " screen even ) , it will also show a similar blank screen . Can anyone suggest where/what to look for to trouble shoot this issue . Let me know if u want any further details . The lab works well and fine when we do a reboot and then the same problem happens after an hour or so ...

    Read the article

  • Software for Automatic Remote FTP Backup

    - by Baez
    I'm looking for software (free or paid) that could perform a weekly automatic backup to an outside server via FTP. I've looked around and all I seem to be finding is either garbage shareware or free tools that are no longer supported. The system will be backing up from a Windows 7 desktop system to a Linux CentOS 5 server. Can someone direct me to a stable, reliable piece of software? This is for business documents so reliability is key.

    Read the article

< Previous Page | 109 110 111 112 113 114 115 116 117 118 119 120  | Next Page >