Search Results

Search found 9067 results on 363 pages for 'big fizzy'.

Page 161/363 | < Previous Page | 157 158 159 160 161 162 163 164 165 166 167 168  | Next Page >

  • Outlook is trying to retrieve data from Microsoft Exchange

    - by Adrian Baker
    Some of my users on my network keep getting a pop up message saying "Outlook is trying to retrieve data from Microsoft Exchange". Microsoft Office 2003 is installed on these computers and when this message appears it freezes every users outlook. I have read some articles, and they say it a problem with 2002 and 2003 version of outlook. So I have changed some users Office to MS Office 2007 but still I have the same problem. This is become a big problem as it is happening often and users are getting annoyed. If anyone can give me some advise on how to solve this problem and what could be causing this problem I would be most grateful.

    Read the article

  • Can't burn 8.1G iso onto 8.4GB DVD - "Media does not have enough free space"

    - by Max Williams
    I'm trying to burn a dvd on a mac with an external (firewire-connected) dvd drive. I'm checking the size of the iso thus: DVD-4:dvd_files macbook$ ls -l /tmp/hybrid.iso -rw-r--r-- 1 macbook wheel 8700884992 Aug 22 10:57 /tmp/hybrid.iso DVD-4:dvd_files macbook$ ls -lh /tmp/hybrid.iso -rw-r--r-- 1 macbook wheel 8.1G Aug 22 10:57 /tmp/hybrid.iso The "human-readable" size is 8.1 Gig but when i try to burn, onto an 8.4G dual-layer dvd, it says "Media does not have enough free space" The definition of a "Gigabyte" according to Wikipedia is 1 billion bytes, so the iso size should actually be 8.7 Gig according to this definition, in which case the disk definitely isn't big enough, and it's just that the -h option to ls is misleading. Is the discrepancy just due to the ls command using a different definition of "G" (eg 1024 Meg aka 1.07 Gig? This comes out as 8.103 which fits what ls is displaying)

    Read the article

  • Multicast File copy with Unicast responses

    - by kirbuchi
    I'm trying to do some multicast big file copies over to remote clients on the other side of a satellite link. The idea is to minimize the amount of traffic going up to the satellite. I tried using uftp without luck. The problem is that, even though we can reach clients via multicast from our central Hub, they aren't able to respond to a multicast address (it's not supported by the return link). As uftp needs to respond to a multicast address in order to report any missing packets I'm out of luck. So does anyone have any recommendations or alternatives I can use to do the trasfers? Any tip or pointer would be appreciated.

    Read the article

  • 4K sectors transition: Why are hard drives moving to 4096 byte sectors, vs. 512 byte sectors?

    - by Chris W. Rea
    I've noticed that some Western Digital hard drives are now sporting 4K sectors, that is, the sectors are larger: 4096 bytes vs. the long-standing standard of 512 bytes. So: What's the big deal with 4K sectors? Is it marketing hype, or a real advantage? Why should somebody building a new PC care, or not, about 4K sectors? Why is this transition taking place now? Why didn't it happen sooner? Are there things to look out for when buying a 4K sector hard drive? e.g. incompatibility? Anything else we should know about 4K sectors?

    Read the article

  • Windows file Sync

    - by Deane Venske
    So I have a big problem at the moment. Trying to find a reliable solution for syncing 2 windows IIS servers. I need to keep the web content imaged on both. Now I have been trying to use Rsync to this point, but unfortunately file permission errors are a nightmare to manage this way. I'm testing out dropbox, but the performance sucks. I'm more familiar with Linux stuff and I've used Rsync in the past but isn't there a native windows solution that will work?

    Read the article

  • SSD on Vmware ESXI 4 (TRIM? Good Idea?)

    - by nextgenneo
    Hi, I just posted about finding bottle necks and have narrowed it down to having way too many VMs on my machine on one 15K SAS drive. I have plenty of cores and plenty of ram. So I am planning on putting 6 VMs on one drive (so 5 drives for 30 VMs). I am thinking of using a 60GB Vertex 2 SSD. Each of my VMs will only need about 6GB of HDD space so this isn't a big deal. My questions are: does ESXI support Trim and do I really need it if I leave 25% of the drive as free space? If I need it should I get a diff drive that handles garbage collection differently? I have a RAID controller w/ write caching. I will still benefit from this? Will this effect my setup differently? Is there anything I need to consider regarding SSD's in virtualized environments. Thanks for any and all help!

    Read the article

  • Running OpenVZ virtual servers within a Xen XCP vritual server? Bad practice?

    - by Damainman
    I have a 1 server with 8GB RAM and 2xQuadcore Processors. It currently has the Xen XCP installed on it, and centos6.2 x64 running on a virtual machine. I have a server control panel software that I want to use and it allows the administration via a web interface for Openvz machines. My questions are: Would this be considered bad practice? Would there be a big performance hit? Should I avoid this all together or am I going about it all wrong? Thank you in advance.

    Read the article

  • In BASH, are wildcard expansions guaranteed to be in order?

    - by ArtB
    Is the expansion of a wildcard in BASH guaranteed to be in alphabetical order? I forced to split a large file into [10Mb pieces][1] so that they can be be accepted by my Mercurial repository. So I was thinking I could use: split -b 10485760 Big.file BigFilePiece. and then in place of: cat BigFile | bigFileProcessor I could do: cat BigFilePiece.* | bigFileProcessor In its place. However, I could not find anywhere that guaranteed that the expansion of the asterisk (aka wildcard, aka '*' ) would always be in alphabetical order so that .aa came before .ab ( as opposed to be timestamp ordering or something like that ). Also, are there any flaws in my plan? How great is the performance cost of cating the file together?

    Read the article

  • Why does internet explorer 8 fail to locate jpg files that other browsers can find?

    - by user278457
    The following URL doesn't display for me in Internet Explorer 8. I even tried compatibility mode and it didn't fix the issue. http://beat.com.au/sites/default/files/images/_DSC5596.jpg It appears just fine in Chrome/Safari/Firefox. I suspect it has something to do with the filename starting with _ but that seems like a fairly big stretch to me. Is this error repeatable on other people's computers? And why on earth would such a strange thing happen anyway?

    Read the article

  • Excel conditional selection?

    - by Andrew
    I think this is a simple question. I have a big table of data points and I want to take a an average of a subset of a single column. For example, if A is "age" and B is "gender," what command could I use to calculate the average age of women in my table? I know I can do this by sorting the table by column B and then only selecting column A values, but I want to build up to much more complicated conditional terms (e.g. if A is 5 and B is 3 and C is 4, then give me the average of D). Thanks!

    Read the article

  • Server installation logging / logbook / diary?

    - by The MYYN
    Are there some ways field-tested ways to keep a kind of logbook for a server? Including: software installations (and de-installations) custom configurations (e.g. of a webserver, ssh daemon, etc.) personal notes The big picture. I am preparing a server and would like to extensively document the state and how it was established over time, so that a new person can easily see, what's going on and why. The setup is not too complicated, but I would like to do it anyway. I once used something like Maintain /etc with mercurial on Debian and it was nice, but I am looking for a little more flexible solution. Addendum: So I am interested in logging and documentation first. In an ideal world however, I would like to have a command, which in a few steps would take me from a bare newly installed unix system to a functional environment with all the components setup and in place by the means of, say an 'executable' log. But that would be very ideal, I imagine.

    Read the article

  • Disable CTRL + ALT + [LETTER] to produce accented variations

    - by Barney
    After an unknown config change, CTRL + ALT + [LETTER] has started producing common accented versions of [LETTER]. I'm not a big fan of this arrangement, seeing as I've memorized all my favourite ALT + [NUMPAD SEQUENCE] references and was used to using CTRL + ALT + [LETTER] for various other application-specific commands in my text editor. The prominent result from my searching suggests that this has something to do with a switch to an 'international keyboard', and says this can be removed in the control panel or toggled by hitting ALT + SHIFT, but I can't get my system to confirm this, and the solutions (or close approximations thereof) don't work. Specifically, I've been to Control Panel\Clock, Language, and Region\Language\Advanced settings and switched the override for default input method from 'language list' to English and I've been to Control Panel\Clock, Language, and Region\Language\Language options and made sure that I only have my one input method (UK). Other than that I'm not quite sure where to look. Any ideas?

    Read the article

  • How can I sync Access databases and keep them up-to-date?

    - by user327472
    I have an Access database on my server. We split it up and use the front-end database for search data and adding new records or reports in local computer. If we update or add a new record, that writes to the back-end of database. I want to use this database in the other building with other servers. Also, those servers have no direct connection. How can I sync both back-end databases to keep the database data up to date? These details may be useful: It's a big amount of data - about 25,750 client records. I guess there are more than 25 tables at 80 MB.

    Read the article

  • Exchange Failover Solution

    - by Dan
    I've been given the task to come up with an exchange solution that will support 200 users total throughout 4 states. 1GB per user. It needs to have a failover solution,The failover must reside in another location. There is an mpls that connects the locations. I am hoping to get recomendations on hardware, software setups. I recently worked with some big name reps and they steered me in the wrong direction and now I'm a few days away from my proposal date with bogus quotes and scrambling for a solution. I used to manage a standalone 2003 exchange server for years and am at a loss now with figuring out a clustering/failover...Any help would be greatly appreciated. thank you

    Read the article

  • Why am I unable to access the network shares on my own computer?

    - by DoomStone
    I am having some problems with media center computer, with Windows 7 on it. The computers IP address is 192.168.1.13, and another computer can without any problem navigate to \\192.168.1.13, but the computer itself can not. Neither \\servername nor \\localhost. If I reboot the computer a couple of times, it will work again, but this is a big problem as my computer cleans my media center library on boot and there, if it can not access \\192.168.1.13, it will think that all the files have been removed and therefore delete them from the database. I really have no idea on where to start with this problem, but if anyone has tried this before and fixed it, I would be very happy, as it would save me from having to reinstall the whole system.

    Read the article

  • TPM had to be reintialized: Does a new recovery password have to be uploaded to AD?

    - by MDMoore313
    Some way some how, a user's machine couldn't get read the bitlocker password off of the TPM chip, and I had to enter the recovery key (stored in AD) to get in. No big deal, but once in the machine, I tried to suspend bitlocker per recovery documentation, and got an error message about the TPM not being initialized. I knew the TPM was on and activated in the BIOS, but Windows still made me reinitialize the TPM chip, and in the process it created a new TPM owner password. I found that odd because it prompted me to save this password or print it (there wasn't an option not to), but it made no reference of a recovery password, nor did it back this password up to AD. After the user took her laptop and left I started thinking that if the TPM password change, does the recovery password change also? If so, that new recovery password will need to be uploaded to AD, but MS' documentation doesn't make that clear, and doesn't back up the new recovery key (if one exists) to AD automatically when the group policy says it must, and from a network standpoint AD is accessible.

    Read the article

  • Seeing double-cursor in 'recursive RDP'

    - by John
    I'm using a Virtual PC VM on my PC to connect to a remote PC using RDP over a VPN. It works just fine except that in the RDP session, I often see a double mouse cursor... I get a standard mouse-arrow and the context-specific cursor such as a caret or hourglass. It's not a big problem but rather annoying and I wondered if there's any way to fix it, or at least understand where it comes from? Is it because I'm effectively running one RDP session from inside another? My host PC runs W7, my local VM runs XP, and the remote PC access using RDP is W7. Thankyou for any help.

    Read the article

  • Moving software RAID to Linux

    - by terman
    I'm using a RAID 1 (mirrored pair) configuration in my Media Center/ NAS system. Currently it's running Windows 8 (yeah, big mistake I know) and I'm regretting it (did it for the games, not worth it). Currently I'm having two software RAID 1s (3TB + 2TB) configured with Storage Spaces and unfortunately formatted with NTFS. Now I would like to switch to Fedora (or maybe Ubuntu if there are advantages) for good. Is there a way that I could continue using the disks as they are without the need to format them with ext or something? I'm glad for every tip. Oh, the system disk is of cause not in a raid configuration.

    Read the article

  • Windows user trying install Git on Solaris

    - by nahab
    Is there simply way to install Git on Solaris as on Windows without installing any side libraries and compiling source files? And if not, why? UPD. Yes I'm looking for single package that will be easy to install. We have ~8 solaris zones using for development those we need simple way to install git fast on they. Installation should be easy because each member of team possibly will be do it and it should be fast because of big count of zones.

    Read the article

  • Linux find/search root partition ONLY?

    - by ~sd-imi
    Say I need to do: find / -name somefile.txt and say root partition / is mounted on /dev/sda5; however, let's say I also have 250GB partitions (/dev/sda6, /dev/sda7) mounted in /media - AND another location that I cannot currently remember. Say, also, that I know the file I'm looking for is on /dev/sda5. Obviously, the above command will also descend in /media and that other directory which represent the big partitions, wasting time in looking for the file in the wrong place. Is there a way to instruct find (or other command) to search only / on /dev/sda5, and NOT to descend to directories if they are on different partitions ? Thanks, Cheers!

    Read the article

  • Not possible to extract tar.gz archive in Centos

    - by Petuga
    I have just uploaded tar.gz archive to my Centos server from my home Windows 7 computer. Before that, it was extracted from .zip archive. When I try to extract it: tar -xvf file.tar.gz gunzip file.tar.gz It always give me error: "not in gzip format" The archive is not corrupted, because it is possible to extract it in Windows with Winrar. The archive is quite big, more than 100 MB. I don't know, what's the problem. Could you help me please? Thanks. Peter

    Read the article

  • VPC on Windows 7 very slow network

    - by Shigg
    I have a Windows 2003 virtual machine which I use for website testing. I've just installed Windows 7 and am using the new version VPC (not xp mode). When I try to copy a file - I need to copy some big databases across - I get a file copy speed of about 20k per sec. Copying from one PC to another on the real network transfers files at 13mb per second. Any ideas what may be causing this? I've turned off differential network compression on win 7. The Virtual HD is on a seperate physical drive to the OS. Running Windows 7 64 bit on a dual xeon with 16gb ram and 10,000 rpm drives. Tried installing VPC 2007 but windows blocks it running saying its not compatable. Many thanks for any ideas.

    Read the article

  • solr administration

    - by devrick0
    Does anyone have any notes for an sysadmin supporting solr? I'm looking for anything that might be useful for monitoring & metrics as well as troubleshooting. Some useful links I have found are: /solr/admin/stats.jsp and /solr/admin/analysis.jsp In the logs I have noticed, other than the query, "hits", "status" and "QTime" values. The documentation on what these mean is sparse at least based on the 100+ websites I have checked. QTime appears to be the query time response in milliseconds. Hits is some form of results but I'm not sure exactly what makes that up and I'm not sure about status. Typically I see status come back as "0" but I have seen other numbers such as "5", so my thoughts that it could be either HTTP status codes or a 0 or 1 (good or bad) methodology isn't accurate. All of the documentation I have come across is intended for developers. Any sysadmin-centric documentation would be a big help.

    Read the article

  • Make my web-server traffic go through proxy?

    - by Eli
    I have a question that may or may not be possible. Basically Comcast isn't going to let me host a website on their ISP unless I have a business account. So, I spoke with my uncle who is big into networking and he told me to host my website through a proxy so Comcast cannot associate the website IP with my IP. I have purchases a proxy from proxy-hub.com, but I cannot seem to figure out what to do next. I may be approaching this totally wrong and I may need to create my own proxy server. Anybody have a clue? Thanks.

    Read the article

  • Distributed filesystem across a slow link

    - by Jeff Ferland
    I have an image in my head where a link is too slow to realize the real-time transfer of files, but fast enough to catch up every day. What I'd like to see is a master <- master setup where when I write a file to Server A, the metadata will transfer to Server B immediately and the file will transfer at idle or immediately when Server B's client tries to read the file before Server A has sent it. It seems that there are many filesystems which can perform well over fast links, but I don't know of any that do well with a big bottle neck and a few hours of latency.

    Read the article

< Previous Page | 157 158 159 160 161 162 163 164 165 166 167 168  | Next Page >