Search Results

Search found 12887 results on 516 pages for 'small jam'.

Page 228/516 | < Previous Page | 224 225 226 227 228 229 230 231 232 233 234 235  | Next Page >

  • How can I troubleshoot SSD disconnection problems on Windows 7?

    - by 0xC0000022L
    I have a small SSD (Transcend StoreJet TS256GESD200K) which I am using on several computers. The drive is formatted with NTFS and recently I started noticing disconnects. Normally one probably wouldn't notice (you don't even get the notification sound in Windows when it disconnects), but since I use USBDLM and run a VM from that SSD, I get to see this first-hand. How could I best figure out whether the problem is the hardware (SSD, USB socket, USB cable or something in the PC) or software? In short: how can I locate the source of the disruption so that I can work on removing the problem? Side-note: SMART status for the SSD is clean.

    Read the article

  • Best technique for reusing a Windows system image across configurations

    - by Martin Wiboe
    We are a small company that provides solutions for ventilation systems. Part of the solution is a "controller" which communicates with the ventilation equipment. These controllers are simply Dell computers that come with our Windows 7 system image on them and sometimes some special hardware. We typically do a batch of 10 controllers at a time. We have been using Norton Ghost to apply the system image, but this process breaks because Dell changes the system configuration often, and our Windows image now does not contain the correct drivers. This is especially a problem when they change the RAID controller. To improve this, I see 2 options: use some kind of virtualization and install a hypervisor on each PC. This would solve the driver problem, but probably cause trouble with our special hardware. use some method of adding the proper drivers to our Windows image in offline mode. I haven't got much experience in either of these approaches. How would you solve our problem?

    Read the article

  • Can KVM roll back changes to Virtual Disks automatically?

    - by Cygon
    I'm currently using VirtualBox on my Linux server to run a small Windows guest OS. I've configured its main virtual hard drive as what VBox calls "Immutable" - meaning that any changes to it are written into a differencing image that is discarded when the system reboots. Can KVM do something similar? I've read about snapshots via "savevm", "loadvm" but I believe that's saved states, not differencing images. What I ultimately want is a VM with two drives: one reverts on each reboot, one keeps its changes. Ideally, the unchangeable drive image should be stored with only read access granted to the user running KVM.

    Read the article

  • Looking for Mini itx capable of running Photoshop + Illustrator (cs4 or higher)

    - by drozzy
    Ok, so I want to build a small pc for my gf that she can take with her instead of the crappy laptop that she has. Overall I think the complete system should not weight more than 10 lbs. The requirements are to run at least 4 applications simultaneously, and be able to switch between them with no problems: Photoshop Illustrator Word editor Browser Should be able to handle 1920x1200 resolution. I am currently looking at LGA775 socket as I can just transfer my desktop cpu Q6600 to it. Currently deciding between DQ45EK and DG41TX, but any other suggestions are welcome. So I am thinking something along the lines of: MINI-BOX M350 Case with 90 watt psu Q6600 cpu (my desktop cpu) 2x2GB kingston ram (or similar) Video? Need external or built-in G45 on DQ45EK will do? My primary concern is whether the 90WATT is sufficient of the Q6600? Thanks

    Read the article

  • How can I increase space on the Filesystem linux?

    - by xtrimsky
    I am renting a dedicated server with Parallel Plesk on it (which I hate and I try to use command line). I have a filesystem that is full,"df -H" prints this: Filesystem Size Used Avail Use% Mounted on /dev/md1 4.0G 4.0G 361k 100% / /dev/mapper/vg00-usr 4.3G 1.4G 3.0G 32% /usr /dev/mapper/vg00-var 4.3G 2.8G 1.6G 64% /var /dev/mapper/vg00-home 4.3G 4.4M 4.3G 1% /home none 1.1G 24M 1.1G 3% /tmp tmpfs 1.1G 0 1.1G 0% /usr/local/psa/handlers/before-local tmpfs 1.1G 0 1.1G 0% /usr/local/psa/handlers/before-queue tmpfs 1.1G 0 1.1G 0% /usr/local/psa/handlers/before-remote tmpfs 1.1G 0 1.1G 0% /usr/local/psa/handlers/info tmpfs 1.1G 0 1.1G 0% /usr/local/psa/handlers/spool The server I'm renting has 1TB of hard drive. Why are these so small, how can I increase my storage ? (I'm pretty beginner with Linux). Thank you

    Read the article

  • Windows 7: Server disappearing from Network window

    - by David Veeneman
    I am running a small (three computer) network on a Windows Server 2003 domain. All of the computers were just upgraded to Windows 7. Since the upgrade, the icon for the domain server periodically disappears from the Network window on all three computers. It doesn't happen at the same time on all computers--the icon typically disappears from one machine, but not the others. And it's not always the same computer. The network connection is still there--I can type \\myserver and get to the network shares. It's simply the icon that's disappearing. Has this happened to anyone else? Any suggestions for a fix? Thanks in advance.

    Read the article

  • IIS FTP service - download timeouts and restarts getting the data twice

    - by accel229
    We have an IIS FTP site on a Windows Server 2003 x64 machine. Application Layer Gateway service is disabled (so http://support.microsoft.com/kb/931130 does not apply). Windows Firewall service is disabled as well. Connection timeout for the FTP site (there is only one) is set to 1,200 seconds = 20 minutes. An external client can connect to the site, list directory contents and download small files. When a client attempts to download a large file (eg, if the download continues for 3 minutes, which is still under 20 minutes, but relatively long), the server sends all data, then the connection times out, the client issues REST / RETR commands attempting to restart the download since after the last byte (which I believe should succeed and receive exactly 0 bytes), and the server behaves as if the client tried to restart after byte 0, that is, it sends the entire file all over. Any ideas on how to fix this?

    Read the article

  • Shrink system volume, create new partition, & enlarge current partitions w/o losing data?

    - by studiohack
    My machine has a 300 GB hard drive. I'm planning to reinstall Windows 7 on this machine soon. It has three partitions: a System partition, and two equal data partitions for particular types of data (music, photos, docs, etc). These data partitions are too small for the data they hold, and the system partition is large enough to be shrunk. I want to create four partitions in preparation for the fresh reinstall, a System partition, and three data partitions of varying sizes (40, 50, & 60 GB respectively). My question: Can I shrink the System volume, create another partition, and make the existing two data partitions bigger without losing any data that is currently on the hard drive? Or is it better to format the whole thing, and create partitions after (thus removing the current OS) with a partitioning tool booting from CD? Can I reformat and create partitions with the Windows 7 set up DVD?

    Read the article

  • How to deny access to disabled AD accounts via kerberos in pam_krb5?

    - by Phil
    I have a working AD/Linux/LDAP/KRB5 directory and authentication setup, with one small problem. When an account is disabled, SSH publickey authentication still allows user login. It's clear that kerberos clients can identify a disabled account, as kinit and kpasswd return "Clients credentials have been revoked" with no further password / interaction. Can PAM be configured (with "UsePAM yes" in sshd_config) to disallow logins for disabled accounts, where authentication is done by publickey? This doesn't seem to work: account [default=bad success=ok user_unknown=ignore] pam_krb5.so Please don't introduce winbind in your answer - we don't use it.

    Read the article

  • Shared storage for web cluster

    - by user52475
    Hi all! Have a big question about shared/clustered/distributed file system for storage. It will shared storage for shared web hosting (web files + maildir) and OpenVZ containers storage . Have any one working example of such system? The options are: Lustre GFS1/GFS2 - GFS2 - as I understand is EXPERIMENTAL... NFS This 3 systems which I consider for shared storage. Now I have storage with HW RAID 10 - 1TB. NFS - As I know there will be problem with locking? GFS/Lustre - problems when there will be a lot of small files , what is typical for hosting environment and problems with maildir.

    Read the article

  • How do I create a free server-side database to be accessed via Windows forms and/or browser?

    - by NoCatharsis
    I have no formal education in databasing or programming, but I've learned enough SQL, C++, and C# to at least get started setting up a small database on my company's server. Using MS SQL Server 2008 R2, I have created the database and set up columns with proper data types. However, there seems to be a lot of tweaks and details that are way over my head. Since I would like these data to be accessible to the other 7 or 8 people in my office (preferably via web browser), I'm wondering whether this is the best setup for my situation. The other option I've read about is a LAMPP server, which I assume is the competing free option to Microsoft's Express packages. I know nothing of LAMP servers except from the articles I've read on how to set them up (and I believe I even saw a detailed tutorial somewhere). To summarize, my question is this: Which of these (or any other) server setups would best suit my purposes, keeping in mind that I'm a true novice (but willing to learn), and would like to keep it free until I get more experience?

    Read the article

  • How do I replace a harddrive that is in a two-way mirror storage space on Windows 8?

    - by Jon
    I have a storage space in Windows 8 doing a two-way mirror on three harddrives. The sizes are 297GB, 189GB, and 70GB. I would like to replace the 70GB HD with a larger one. My thought was to remove that drive from the space via the Storage Space control panel, shutdown, replace HD with bigger drive, reboot, add new HD to the storage space. I can't find any options to remove a HD from a storage space in the control panel. Should I just shutdown and swap out the small drive or is there another process for safely replacing the old HD? (By the way, the old HD is still operational.)

    Read the article

  • How To Create An FTP User That Has Permission To EVERYTHING

    - by Serg
    I've spent the last two hours trying to create an FTP user so I can transfer some files over to my Wordpress blog folder. /var/www/sergiotapia.me I'm using vsftpd and Ubuntu 12.04 for my FTP server and I've read tons of documentation, none of which seem to work. I still cannot log in with the FTP user, let alone test if I even have the read/write file permissions. Can a Linux guru here, help me out with a small step by step? I'm comfortable with the terminal and nano, so that's not an issue - I'll SSH into my box. Just tell me what to do and what commands to run. Specifically, this user needs to have read and write access to the /var/ folder and anything within it. I want to have 1 user that can do whatever the heck he wants on my Ubuntu 12.04 VPS machine.

    Read the article

  • script to automatically test if a web site is available

    - by Xoundboy
    I'm a lone web developer with my own Centos VPS hosting a few small web sites for my clients. Today I discovered my httpd service had stopped (for no apparent reason - but that's another thread). I restarted it but now I need to find a way that I can be notified by email and/or SMS if it happens again - I don't like it when my client rings me to tell me their web site doesn't work! I know there are probably many different possibilities, including server monitoring software. I think all I really need is a script that I can run as a cron job from my dev host (which is permanently running in my office) that attempts to load a page from my production server and if it doesn't load within say 30 seconds then it sends me an email or SMS. I'm pretty rubbish at shell scripting, hence this question. Any suggestions would be gratefully appreciated, thanks to all you clever sysadmin guys and girls out there :)

    Read the article

  • Sync Banshee library data.

    - by Dom
    I use Banshee to organise my music, I particularly like its scoring system and I have smart playlists based on it. However, I have two versions of my music library, one on each of my computers. As one of the computers is small I only have a favourite set of songs on that computer rather than my whole collection. The computers are not on a local network, but I do use Ubuntu One for file sharing between them. Is there any way I can synchronise song data (play count, score, skip count ...) and playlist data (including smart playlists that include songs based on this data) between the two computers? This would only be relevant of course for the songs that exist on both computers, the songs that exist on only one would need to be ignored. I did consider putting the library data file (I think it is .xml but I'm not sure) into the shared file and creating a symbolic link to it, but then I wouldn't be able to have a different set of songs on each computer. Thank you.

    Read the article

  • Do I need a VPN to secure communication over a T1 line?

    - by Seth
    I have a dedicated T1 line that runs between my office and my data center. Both ends have public IP addresses. On both ends, we have a T1 routers which connect to SonicWall firewalls. The SonicWalls do a site-to-site VPN and handle the network translation, so the computers on the office network (10.0.100.x) can access the servers in the rack (10.0.103.x). So the question: can I just add a static route to the SonicWalls so each network can access each other with out the VPN? Are there security problems (such as, someone else adding the appropriate static route and being able to access either the office or the datacenter)? Is there another / better way to do it? The reason I'm looking at this is because the T1 is already a pretty small pipe, and having the VPN overhead makes connectivity really slow.

    Read the article

  • Hosting a web site at home: ISP Blocking port 80?

    - by tombull89
    Hello, I presume this is a better place to put this rather than server fault. I'm interested in setting up a small site to host at home as a "proof of concept" exercise, i.e. to prrove that I know how to do it. I've got a (virtual) server 2003 machine with a site on it, all configured with port forwarding through to 80 on my server. I have a Belkin F5D7634 which I have put my DYNDNS details in but when I try to go to my DYNDNS address it comes up with the page cannot be displayed. My ISP is Carphone Warehouse/AOL and I've been unable to find any information if they block port 80. If they do, can anybody reccomend a home provider that does not block port 80? Regards, Tom.

    Read the article

  • Should I limit end-user gigabit ports to avoid saturating uplink/trunk connections?

    - by Joel Coel
    We have a campus with 16 buildings and older 850nm 1Gbps fiber links between the buildings, that all come to a core switch for our servers that also uses 1Gbps ports. We're finally starting to replace our aging 10/100 end-user switches, and much of what we're looking at are 1 Gbps units. My question is, since the trunk/uplink lines are still 1Gbps, if I were to install 1 Gbps switches for end users, should I limit the ports to 100Mbps until I can also upgrade the trunks to avoid allowing a bad-behaving host to saturate a trunk line (since we're a college, we have plenty of mis-behaving hosts) and thereby create a DoS situation for a building, or will TCP congestion control typically take care of that for me? What if we have a lot of UDP traffic (games, video chats, even a small amount of bittorrent)?

    Read the article

  • Ubuntu based VPN server behind NAT router?

    - by maartenn88
    I have a small intranet- and file-server inside our company, which can only be reached from inside our own network. I'd like to be able to reach the intranet and files on the server from outside the network, however I don't really like the idea of completely opening security by forwarding the ports on our NAT-router, so I'd prefer to enable a VPN possibility for this. So I have a router, which is the DHCP server and has NAT function, and I have an Ubuntu Server with the files and intranet on it. How do I configure these two so that I can make a VPN connection with my network in order to reach the server? thanks in advance for any answer!

    Read the article

  • Minimum size of a boot partition on debian

    - by zebonaut
    I'm setting up an old box with Debian. First etch (4.0), because this is the last version that still had boot floppies, then the box is to be upgraded to lenny (5.0) and squeeze (6.0). Therefore, I will end up having a a couple of different kernel versions in the boot partition. If I don't want to be wasteful and if I end up needing a separate boot partition, how large should it be? I've used 10 MB long ago, but that was woody, and only one kernel in the boot partition, and this seems to be too small for what I want to do now.

    Read the article

  • How to push to github from a server account with multiple users?

    - by kirdie
    We have a web server which contains a web application stored as a github project. Now all of us can push from our local machines to github and then pull on the server but sometimes we want to make small changes and immediately see the effect so it would be great to be able to push at the server too. Now I created an ssh key for the server but I don't want to add the servers ssh key to my github account because then all github actions done from the server are counted to my account. Is it possible to add the ssh key to the github web application project without creating a new user for the server and what is the best practice for this situation? I also don't want to copy my private key to the server obviously.

    Read the article

  • How to know if my nginx is in good health?

    - by Howard
    I am running a nginx on EC2 (m1.small) for SSL termination. I am using 2 workers on Ubuntu, with latest nginx (stable), the network throughput is around 2Mbps and system load average is around 2 to 3. I am wondering if this system is in good health for now, e.g. what is the queue length (I know nginx can handle a lot of concurrent request, but I mean before the request is being served, how many of them need to wait before being served) what is the average queue time for a given request to be served. I want to know because if my nginx is cpu bounded (e.g. due to SSL), I will need to upgrade to a faster instance. My current nginx status Active connections: 4076 server accepts handled requests 90664283 90664283 104117012 Reading: 525 Writing: 81 Waiting: 3470

    Read the article

  • Force ntpd to make changes in smaller steps

    - by David Wolever
    The NTP documentation says: Under ordinariy conditions, ntpd adjusts the clock in small steps so that the timescale is effectively continuous and without discontinuities - http://doc.ntp.org/4.1.0/ntpd.htm However, this is not at all what I have noticed in practice. If I manually change the system time backwards or forwards 5 or 10 seconds then start ntpd, I notice that it adjusts the clock in one shot. For example, with this code: #!/usr/bin/env python import time last = time.time() while True: time.sleep(1) print time.time() - last last = time.time() When I first change the time, I'll notice something like: 1.00194311142 8.29711604118 1.0010509491 Then when I start NTPd, I'll see something like: 1.00194311142 -8.117301941 1.0010509491 Is there any way to force ntpd to make the adjustments in smaller steps?

    Read the article

  • How to connect two subnet with Windows Server?

    - by 9dan
    Hello, I have some work to do in a small office. This office has two separate Internet connections. One is a kind of DSL line and connected through the Internet sharing router (DHCP). One is a leased line with public IP adresses (No DHCP). This leased line provide only 30 IP addresses and the gateway is provided by the ISP. Some workers use private IP provided by the router, some workers use manually setup public IP. There is a Windows 2008 R2 server with two NICs, connected to each lines. I want to connect/merge(?)/bridge(?) two subnet with this server so that users from the different subnet can access each other. How can I accomplish this? Something like, add gateway function to this server and public IP users change the gateway to point this server.. It it possible ? Sorry for my rambling. Thanks in advance.

    Read the article

  • Apple file sharing: bind to a specific interface

    - by Cesar
    My customer have an office small office with just a wifi router. He use this router for internet connectivity and file transfer operations between the desktops. Recently the file transfer activity between desktop (all osx based) is increased a lot so he bought a switch (no connected to the router, too far away) for transfer the file over the cable instead over the wifi. Problem: How to bind the file sharing service just to the Ethernet interface and exclude the wifi interface ? (actually the service binds to the wifi automatically and there are no options about the interface binding)

    Read the article

< Previous Page | 224 225 226 227 228 229 230 231 232 233 234 235  | Next Page >