Search Results

Search found 17913 results on 717 pages for 'old school rules'.

Page 385/717 | < Previous Page | 381 382 383 384 385 386 387 388 389 390 391 392  | Next Page >

  • How to make Exchange 2010 use In-Reply-To and References headers as well as Thread-Index?

    - by Paul Wagland
    The problem that I have is simple. We have recently upgraded from Exchange 2003 to Exchange 2010. Everything went fine, and there are very few complaints. In Exchange 2003, some of our OWA users liked the Threading view, these users love the conversations view, especially with the cross-mailbox threading. The problem is that some of these users are now complaining that old e-mails that they got from systems like bugzilla were migrated across correctly threaded, but the new e-mails are not being correctly threaded. If I look at the mail source using an IMAP client then I can see that all of the mails that are turning up as being threaded have the (Microsoft specific) Thread-Index header, and the mails that are not getting grouped do not have this header. The question is, how can I make the webmail client respect the normal threading? That is, is there a way to Make the OWA client use the standard In-Reply-To and References headers, or Make Exchange generate the Thread-Index headers for Outlook and OWA to use?

    Read the article

  • Increasing System Menu Font Size for OS X Lion

    - by JW01
    My father-in-law recently picked up an iMac (with OS X 10.7 - Lion) and loves it. However, his eyesight isn't that great and would like to find a way to increase the font size of the system menu bar like he could on his old Windows system. I've been unable to find anything in the system configuration that might work. I've heard that the TinkerTool might be able to do it, but I'm a little hesitant to monkey around with his system - I don't want to introduce system stability issues. Question: Has anyone been able to find any good way to increase the system menu font size?

    Read the article

  • How to get headphones and speakers working at the same time?

    - by Borek
    I have a Gigabyte P55-UD3 motherboard which comes with a Realtek sound card and audio header for the front panel. My headset is permanently connected to the front panel (mic in, headphones in) and I also have speakers connected at the rear of the computer. What I want to achieve is that both the headphones and the speakers work at the same time. On my old PC with Realtek, I was able to do this by clicking "Device advanced settings" in the Realtek HD Audio Manager and then clicking "Make front and rear output devices playback two different audio streams simultaneously" and then doing some reassigning of front/rear settings in the Realtek Audio Manager (can't remember the details). But that doesn't seem to be working now - either the speakers play sounds or the headphones, not both at the same time. To pose the question differently, Windows plays the sounds on the "default device". I'd like to make both headphones and the speakers kind of like the default devices.

    Read the article

  • Moving files fails due to privileges but they seem to be OK

    - by joaoc
    I am trying to copy old files from an OS X 10.5.8 to a new external HD. When trying to copy a folder I get the message: The operation cannot be completed because you do not have sufficient privileges for some of the items. I've checked the privileges and they seem ok (read & write for me). What is curious is that the folder is created empty on the target drive and I can then copy the contents from inside the original folder to inside the new folder without changing anything else. This happens with several folders but not all and is making backup a pain. I have to figure out which folder broke the copy, copy its contents to the external disk and then select what hasn't been copied to copy (and eventually stop at some point and repeating the experience)

    Read the article

  • LDAP query on linux against AD returns groups with no members

    - by SethG
    I am using LDAP+kerberos to authenticate against Active Directory on Windows 2003 R2. My krb5.conf and ldap.conf appear to be correct (according to pretty much every sample I found on the 'net). I can login to the host with both password and ssh keys. When I run getent passwd, all my ldap user accounts are listed with all the important attributes. When I run getent group, all the ldap groups and their gid's are listed, but no group members. If I run ldapsearch and filter on any group, the members are all listed with the "member" attribute. So the data is there for the taking, it's just not being parsed properly. It would appear that I simply am using an incorrect mapping in ldap.conf, but I can't see it. I've tried several variations and all give the same result. Here is my current ldap.conf: host <ad-host1-ip> <ad-host2-ip> base dc=my,dc=full,dc=dn uri ldap://<ad-host1> ldap://<ad-host2> ldap_version 3 binddn <mybinddn> bindpw <mybindpw> scope sub bind_policy hard nss_reconnect_tries 3 nss_reconnect_sleeptime 1 nss_reconnect_maxsleeptime 8 nss_reconnect_maxconntries 3 nss_map_objectclass posixAccount User nss_map_objectclass posixGroup Group nss_map_attribute uid sAMAccountName nss_map_attribute gidNumber msSFU30GidNumber nss_map_attribute uidNumber msSFU30UidNumber nss_map_attribute cn cn nss_map_attribute gecos displayName nss_map_attribute homeDirectory msSFU30HomeDirectory nss_map_attribute loginShell msSFU30LoginShell nss_map_attribute uniqueMember member pam_filter objectcategory=User pam_login_attribute sAMAccountName pam_member_attribute member pam_password ad Here's the kicker: this config works 100% fine on a different linux box with a different distro. It does not work on the distro I am planning on switching to. I have installed from source the versions of pam_ldap and nss_ldap on the new box to match the old box, which fixed another problem I was having with this setup. Other relevant info is the original AD box was Windows 2003. It's mirror died a horrible hardware death so I'm trying to add two more 2003-R2 servers to the mirror tree and ultimately drop the old 2003 box. The new R2 boxes appear to have joined the DC forest properly. What do I need to do to get groups working? I've exhausted all the resources I could find and need a different angle. Any input is appreciated. Status update, 7/31/09 I have managed to tweak my config file to get full info from the AD and performance is nice and snappy. I replaced the back-rev'd copies of pam_ldap and nss_ldap with the current ones for the distro I'm using, so it's back to a standard out-of-the-box install. Here's my current config: host <ad-host1-ip> <ad-host2-ip> base dc=my,dc=full,dc=dn uri ldap://<ad-host1> ldap://<ad-host2> ldap_version 3 binddn <mybinddn> bindpw <mybindpw> scope sub bind_policy soft nss_reconnect_tries 3 nss_reconnect_sleeptime 1 nss_reconnect_maxsleeptime 8 nss_reconnect_maxconntries 3 nss_connect_policy oneshot referrals no nss_map_objectclass posixAccount User nss_map_objectclass posixGroup Group nss_map_attribute uid sAMAccountName nss_map_attribute gidNumber msSFU30GidNumber nss_map_attribute uidNumber msSFU30UidNumber nss_map_attribute cn cn nss_map_attribute gecos displayName nss_map_attribute homeDirectory msSFU30HomeDirectory nss_map_attribute loginShell msSFU30LoginShell nss_map_attribute uniqueMember member pam_filter objectcategory=CN=Person,CN=Schema,CN=Configuration,DC=w2k,DC=cis,DC=ksu,DC=edu pam_login_attribute sAMAccountName pam_member_attribute member pam_password ad ssl off tls_checkpeer no sasl_secprops maxssf=0 The remaining problem now is when you run the groups command, not all subscribed groups are listed. Some are (one or two), but not all. Group memberships are still honored, such as file and printer access. getent group foo still shows that the user is a member of group foo. So it appears to be a presentation bug, and does not interfere with normal operation. It also appears that some (I have not determined exactly how many) group searches do not resolve correctly, even though the group is listed. eg, when you run "getent group bar", nothing is returned, but if you run "getent group|grep bar" or "getent group|grep <bar_gid>" you can see that it indeed listed and your group name and gid are correct. This still seems like an LDAP search or mapping error, but I can't figure out what it is. I'm a heckuva lot closer than earlier in the week, but I'd really like to get this last detail ironed out.

    Read the article

  • SharePoint Server 2007 Stops Responding to External Requests

    - by Mountain Deux
    We have a SharePoint2007 server running on a Windows Server 2003 Box. The box itself is quite old but has not experienced any weird issues. Recently SharePoint has begun to show some strange behavior. One of the most annoying things is that once in a while, the sharepoint sever stops responding to requests outside our network. Internally, you can still browse everything that has to do with sharepoint but outside, nothing. If i reset the server it will start responding again. I am just not sure what makes it stop responding to users on the outside in the first place. Resetting the WWW publishing service does no good.

    Read the article

  • Problem closing MDI child window in Terminal Services/Remote Desktop Connection 7.0

    - by Justin Love
    I have one user whose computer just got updated to the 7.0 Remote Desktop Connection. Concurrently, she has started having a problem closing the MDI child windows in an old FoxPro application running on the remote server. We have two different servers, both 2003, running the same application, one locally and one at a remote office. Only the remote office server is giving trouble. It works fine for me, even when logging into her TS account. No other users have complained. The other day the same user experienced an error message (path not found for a path showing a localization placeholder) starting the RDC, fixed by reboot. I suspect she may have had RDC running during the 7.0 upgrade.

    Read the article

  • Remove subdomain from Google Crawler

    - by Walter White
    Hi all, I recently removed a sub-domain from my domain so I just have 1 website to manage. However, if I do a google search, my old domain is still there, I removed the sub-domain well over a week ago and if you try to access the domain directly, you will get an error saying the website can not be found (the records have been deleted). What is the easiest way to remove that sub-domain from google searches since they no longer exist? Shouldn't google see that the domain no longer exists and delete those entries? Walter

    Read the article

  • Modify Okular highlight to automatically copy highlighted text into comment

    - by JDD
    Despite what old SE questions state, the PDF software Okular can now write annotations directly to the PDF. This makes it very useful in conjunction with Docear for annotating academic literature. However, Docear imports annotations from the comments, rather than importing from highlighted text. In Okular, when you highlight text it can then be clicked to reveal a comments bubble, which is empty by default. Copying the highlighted text into the resulting bubble allows it to be imported into Docear, but this is laborious. How can I modify the highlight tool to automatically copy the highlighted text into the resulting comment bubble?

    Read the article

  • Live noise-filter on line-in

    - by Damon Gant
    I'm running the following setup: Xbox 360 is hooked up to my (PC) screen via HDMI/DVI converter. Because the Xbox has no dedicated sound output, except for optical S/PIDF, I'm also using the AV/RCA output, namely just the audio, which is connected to an old stereo, which is then connected to my PCs line-in. I'm now experiencing a some of noise. I'm using one of the standard "Realtek High Definition Audio" cards, which doesn't seem to offer this kind of functionality. Is there a software that will playback audio right off a device while running filters on it? It doesn't have to create a device on its own, I just want to listen to it. Here's a sample: http://puu.sh/1suY6

    Read the article

  • How do I change the location of DivX cache files?

    - by andygrunt
    I recently installed to the latest version of DivX and suddenly found my C drive filling up with the cache files. I tracked it down to: C:\Files\My Videos\DivX Movies\Temporary Downloaded Files My old laptop (running WinXP) only has a small hard drive and any DivX cache files fills it up so I want it to use my D drive where I have a little more room. The trouble is I can't see anywhere in the DivX preferences where I can change the cache location. Can anyone tell me how I can change the location of the DivX cache files?

    Read the article

  • ubuntu 9.10 screen flickering problem on Thinkpad R31 with Intel 83830 Graphics

    - by PA
    I am trying to revitalize an old Thinkpad R31 that has the Intel 82830 graphics and only 256 MB of RAM. I have tried a Xubuntu 9.10 Live CD. After booting the screen blinks so much that it is practically unreadable. I have searched for updated IBM Thinkpad drivers but I only found drivers for Windows on the Thinkpad support web site. EDIT: I have changed the title and the description. It is not a problem of the drivers. It seems to be a problem with the kernel. See my own answer below.

    Read the article

  • Is there a way to change the GTK theme for applications run as superuser on KDE?

    - by Patches
    When I run GTK applications on KDE, they use the QtCurve theme that matches my color and font scheme as configured in the KDE System Settings application. However GTK applications run as superuser use the old default GNOME, regardless of whether I run them with kdesudo, gksudo, or sudo on a terminal. For example, here's gedit run as superuser on top, and under my normal user account on the bottom: Strangely, KDE applications run with kdesudo display the default Oxygen styling but use my settings when run with sudo on a terminal. Is there any way to configure the stying GTK applications use when run as superuser on KDE?

    Read the article

  • Install Office 2003 professional over Office 2003 Basic?

    - by Paul
    I have a problem with one of our PCs. I've inherited this PC and it's always had MS Office Basic on it. The time has come to install our Access application on it, so I'm trying to install Office Professional 2003. The problem I have is the installer says the license key is invalid. I can only assume this is down to the fact that the PC used to have Office Basic installed. I've deleted the registration keys for the OFFICE11 section of the registry, but am still getting the problem. Any suggestions as to how to kill any trace of the old install would be appreciated - it's no longer showing up in the add/remove programs window.

    Read the article

  • Choosing a very basic Linux distro with minimal boot time?

    - by user30961
    I've had Ubuntu for a year now, like it but it doesn't match my needed configuration for my old laptop. So... I am using Linux mainly for webapps so I don't care much about user interface. What I need is a distro which is small enough to install on a SD memory card (or USB) since I am removing the hard drive, all my data is in the cloud. I also want it to have as short boot time as possible. I tried Puppy Linux and it seems ok but it had preinstalled looots af small apps I don't need, I would really like to have a distro with minimal preinstallations and add them on my own. So what are my choices and why would you recommend these?

    Read the article

  • Linux SW Raid: whole disk or per-partition?

    - by Steve Pomeroy
    I have inherited a machine which has 2 physical disks and uses Linux SW RAID(1). Both disks are partitioned and are are all individual arrays (/dev/md0, /dev/md6, etc.). Those arrays are then mounted (/boot, /home, etc. even /tmp). As RAID is designed to mitigate physical failures, is there any reason why one would use this technique over whole-disk arrays that are then partitioned (perhaps using LVM)? This seems prone to more potential issues, but may have some special properties that I haven't been able to glean. I'm planning on moving this setup to: disks?SWRAID(1)?LVM as I'll be making multiple VMs out of the one machine, but wanted to make sure I knew what I was doing when I got rid of the old setup.

    Read the article

  • solved: puppet master REST API returns 403 when running under passenger works when master runs from command line

    - by Anadi Misra
    I am using the standard auth.conf provided in puppet install for the puppet master which is running through passenger under Nginx. However for most of the catalog, files and certitifcate request I get a 403 response. ### Authenticated paths - these apply only when the client ### has a valid certificate and is thus authenticated # allow nodes to retrieve their own catalog path ~ ^/catalog/([^/]+)$ method find allow $1 # allow nodes to retrieve their own node definition path ~ ^/node/([^/]+)$ method find allow $1 # allow all nodes to access the certificates services path ~ ^/certificate_revocation_list/ca method find allow * # allow all nodes to store their reports path /report method save allow * # unconditionally allow access to all file services # which means in practice that fileserver.conf will # still be used path /file allow * ### Unauthenticated ACL, for clients for which the current master doesn't ### have a valid certificate; we allow authenticated users, too, because ### there isn't a great harm in letting that request through. # allow access to the master CA path /certificate/ca auth any method find allow * path /certificate/ auth any method find allow * path /certificate_request auth any method find, save allow * path /facts auth any method find, search allow * # this one is not stricly necessary, but it has the merit # of showing the default policy, which is deny everything else path / auth any Puppet master however does not seems to be following this as I get this error on client [amisr1@blramisr195602 ~]$ sudo puppet agent --no-daemonize --verbose --server bangvmpllda02.XXXXX.com [sudo] password for amisr1: Starting Puppet client version 3.0.1 Warning: Unable to fetch my node definition, but the agent run will continue: Warning: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /certificate_revocation_list/ca [find] at :110 Info: Retrieving plugin Error: /File[/var/lib/puppet/lib]: Failed to generate additional resources using 'eval_generate: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /file_metadata/plugins [search] at :110 Error: /File[/var/lib/puppet/lib]: Could not evaluate: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /file_metadata/plugins [find] at :110 Could not retrieve file metadata for puppet://devops.XXXXX.com/plugins: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /file_metadata/plugins [find] at :110 Error: Could not retrieve catalog from remote server: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /catalog/blramisr195602.XXXXX.com [find] at :110 Using cached catalog Error: Could not retrieve catalog; skipping run Error: Could not send report: Error 403 on SERVER: Forbidden request: XX.XXX.XX.XX(XX.XXX.XX.XX) access to /report/blramisr195602.XXXXX.com [save] at :110 and the server logs show XX.XXX.XX.XX - - [10/Dec/2012:14:46:52 +0530] "GET /production/certificate_revocation_list/ca? HTTP/1.1" 403 102 "-" "Ruby" XX.XXX.XX.XX - - [10/Dec/2012:14:46:52 +0530] "GET /production/file_metadatas/plugins?links=manage&recurse=true&&ignore=---+%0A++-+%22.svn%22%0A++-+CVS%0A++-+%22.git%22&checksum_type=md5 HTTP/1.1" 403 95 "-" "Ruby" XX.XXX.XX.XX - - [10/Dec/2012:14:46:52 +0530] "GET /production/file_metadata/plugins? HTTP/1.1" 403 93 "-" "Ruby" XX.XXX.XX.XX - - [10/Dec/2012:14:46:53 +0530] "POST /production/catalog/blramisr195602.XXXXX.com HTTP/1.1" 403 106 "-" "Ruby" XX.XXX.XX.XX - - [10/Dec/2012:14:46:53 +0530] "PUT /production/report/blramisr195602.XXXXX.com HTTP/1.1" 403 105 "-" "Ruby" thefile server conf file is as follows (and goin by what they say on puppet site, It is better to regulate access in auth.conf for reaching file server and then allow file server to server all) [files] path /apps/puppet/files allow * [private] path /apps/puppet/private/%H allow * [modules] allow * I am using server and client version 3 Nginx has been compiled using the following options nginx version: nginx/1.3.9 built by gcc 4.4.6 20120305 (Red Hat 4.4.6-4) (GCC) TLS SNI support enabled configure arguments: --prefix=/apps/nginx --conf-path=/apps/nginx/nginx.conf --pid-path=/apps/nginx/run/nginx.pid --error-log-path=/apps/nginx/logs/error.log --http-log-path=/apps/nginx/logs/access.log --with-http_ssl_module --with-http_gzip_static_module --add-module=/usr/lib/ruby/gems/1.8/gems/passenger-3.0.18/ext/nginx --add-module=/apps/Downloads/nginx/nginx-auth-ldap-master/ and the standard nginx puppet master conf server { ssl on; listen 8140 ssl; server_name _; passenger_enabled on; passenger_set_cgi_param HTTP_X_CLIENT_DN $ssl_client_s_dn; passenger_set_cgi_param HTTP_X_CLIENT_VERIFY $ssl_client_verify; passenger_min_instances 5; access_log logs/puppet_access.log; error_log logs/puppet_error.log; root /apps/nginx/html/rack/public; ssl_certificate /var/lib/puppet/ssl/certs/bangvmpllda02.XXXXXX.com.pem; ssl_certificate_key /var/lib/puppet/ssl/private_keys/bangvmpllda02.XXXXXX.com.pem; ssl_crl /var/lib/puppet/ssl/ca/ca_crl.pem; ssl_client_certificate /var/lib/puppet/ssl/certs/ca.pem; ssl_ciphers SSLv2:-LOW:-EXPORT:RC4+RSA; ssl_prefer_server_ciphers on; ssl_verify_client optional; ssl_verify_depth 1; ssl_session_cache shared:SSL:128m; ssl_session_timeout 5m; } Puppet is picking up the correct settings from the files mentioned because config print command points to /etc/puppet [amisr1@bangvmpllDA02 puppet]$ sudo puppet config print | grep conf async_storeconfigs = false authconfig = /etc/puppet/namespaceauth.conf autosign = /etc/puppet/autosign.conf catalog_cache_terminus = store_configs confdir = /etc/puppet config = /etc/puppet/puppet.conf config_file_name = puppet.conf config_version = "" configprint = all configtimeout = 120 dblocation = /var/lib/puppet/state/clientconfigs.sqlite3 deviceconfig = /etc/puppet/device.conf fileserverconfig = /etc/puppet/fileserver.conf genconfig = false hiera_config = /etc/puppet/hiera.yaml localconfig = /var/lib/puppet/state/localconfig name = config rest_authconfig = /etc/puppet/auth.conf storeconfigs = true storeconfigs_backend = puppetdb tagmap = /etc/puppet/tagmail.conf thin_storeconfigs = false I checked the firewall rules on this VM; 80, 443, 8140, 3000 are allowed. Do I still have to tweak any specifics to auth.conf for getting this to work? Update I added verbose logging to the puppet master and restarted nginx; here's the additional info I see in logs Mon Dec 10 18:19:15 +0530 2012 Puppet (err): Could not resolve 10.209.47.31: no name for 10.209.47.31 Mon Dec 10 18:19:15 +0530 2012 access[/] (info): defaulting to no access for 10.209.47.31 Mon Dec 10 18:19:15 +0530 2012 Puppet (warning): Denying access: Forbidden request: 10.209.47.31(10.209.47.31) access to /file_metadata/plugins [find] at :111 Mon Dec 10 18:19:15 +0530 2012 Puppet (err): Forbidden request: 10.209.47.31(10.209.47.31) access to /file_metadata/plugins [find] at :111 10.209.47.31 - - [10/Dec/2012:18:19:15 +0530] "GET /production/file_metadata/plugins? HTTP/1.1" 403 93 "-" "Ruby" On the agent machine facter fqdn and hostname both return a fully qualified host name [amisr1@blramisr195602 ~]$ sudo facter fqdn blramisr195602.XXXXXXX.com I then updated the agent configuration to add dns_alt_names = 10.209.47.31 cleaned all certificates on master and agent and regenerated the certificates and signed them on master using the option --allow-dns-alt-names [amisr1@bangvmpllDA02 ~]$ sudo puppet cert sign blramisr195602.XXXXXX.com Error: CSR 'blramisr195602.XXXXXX.com' contains subject alternative names (DNS:10.209.47.31, DNS:blramisr195602.XXXXXX.com), which are disallowed. Use `puppet cert --allow-dns-alt-names sign blramisr195602.XXXXXX.com` to sign this request. [amisr1@bangvmpllDA02 ~]$ sudo puppet cert --allow-dns-alt-names sign blramisr195602.XXXXXX.com Signed certificate request for blramisr195602.XXXXXX.com Removing file Puppet::SSL::CertificateRequest blramisr195602.XXXXXX.com at '/var/lib/puppet/ssl/ca/requests/blramisr195602.XXXXXX.com.pem' however, that doesn't help either; I get same errors as before. Not sure why in the logs it shows comparing access rules by IP and not hostname. Is there any Nginx configuration to change this behavior?

    Read the article

  • Checkpoint FW-1disk full

    - by krugger
    Hi, The logging filled the disk of the managment node. This cause the firewall nodes to start logging locally. After deleting some old logs, restarting the managment node logging is still being done locally at the firewall nodes. I have already done a fw fetchlogs on all firewall nodes to get the local log entries. How can I tell the firewall nodes that they should once again connect to the managment node? My syslog_servers show doesn't show any syslog servers. But the documentation mentions that this is for additional logging and not for logging between firewall nodes and the managment node.

    Read the article

  • Dreamweaver Delete Space to Word, Not Word

    - by Don
    There's a built in DW keyboard shortcut (Ctrl + Del) that deletes up to AND INCLUDING the first word to the right. I used to use the ColdFusion Studio app for coding and it would just remove the space UP TO the word (left the word or bracket, or whatever alone.) Any DW users know if this is a setting that can be changed? I'm really used to the old behavior and keep deleting the first word, hitting Ctrl + Z to put it back and then having to manually delete all the spaces to leave the word... Hoping one of you geniuses can help! Thanks, D.

    Read the article

  • Blank white desktop icons on new Windows 7 install

    - by SiegeX
    I just did a fresh install of Windows 7 professional on my father's Toshiba Satellite laptop and everything went smoothly except for the fact that certain desktop shortcuts appear as a blank white page. I've tried: going in and changing the icon to many different things reloading the icon cache by deleting the IconCache.db file uninstalling his Free AVG as he heard the virus scanner might be preventing it (it wasn't and he put it back.) The one thing these icons have in common is that they are shortcuts to some very old DOS executables. One of them is Word Star 2000 to give you an idea. Does anybody have any other suggestions besides what we've tried?

    Read the article

  • Computer starts, POSTs normal, no video

    - by Annath
    So, I recently tried to start up an old computer of mine. It would not power on, so I replaced the PSU. When I powered it on, all the fans spun up and the POST beeps indicated a normal startup. The problem was there was no video output. Every couple of minutes, I hear the post code again. I think it is restarting after POST, but without video I have no idea why. Does anyone have an idea as to why this is happening, and how to fix it? EDIT: The video card is a PCI-e card. There is no integrated graphics on the motherboard. If I remove the video card, the POST indicates a missing card. When I put it back, it goes back to normal, so I know that it recognizes that the card exists.

    Read the article

  • Used HDD/ran DiskSmartView/40,000 Power-on-hours?? should i trust it w/ my data, or take it back and bitch?

    - by David Lindsay
    I just bought a used hard drive from a University Surplus Store. Decided to run DiskSmartView to make sure it wasn't ready to fail. 40,000 power-on-hours I don't know if I feel like trusting my data to something that used. I really dont know if thats unreasonably old, but when i compare it to the POH reading i get when testing my other hdds its more than 3x older (my others have 2110 hours, 6150 hours, etc.. It's a Western Digital, so that gives me a little bit of hope(WDC WD4000KD-00NAB0). I could sure use someone else's opinion here. Thanks, DAVE

    Read the article

  • preparing to migrate MOSS 2007 SP1 content into new server with MOSS 2007 SP2 with the same server n

    - by Albert Widjaja
    To All Experts here, I'd like to perform a MOSS server content migration from the existing single instance server (SQL Server 2008, Windows Server 2008, SharePoint 2007 SP1) all in one and by maintaining the server name only, here's what I had in my migration plan document outline: I have created the new Windows Server 2008 R2 + SQL Server 2008 SP1 and MOSS 2007 SP2 after that: Backup MOSS_ContentDB from SQL Server 2008 SSMS (does this enough to cover all top level sites and its content in the library and doc. repository ?) Backup all top level sites from CA site using the CA Sites backup and restore tools. turn off the old MOSSDEV01 (current server) rename the existing server (TempServ01) into MOSSDEV01 (so that the user bookmark and other link inside MOSS site still working) perform restore of the DB restore the site collection and its content from the backup is that the correct way to do it ? I haven't execute the server configuration wizard to create the same port number of the SSP, CA and the SQL Server DB yet. Any idea and suggestion would be greatly appreciated Thanks.

    Read the article

  • When i log on then i get annyoing cmd windows in windows7

    - by user191542
    I am using windows 7 on my office PC and that is linked with Active directory. I dont't know what happend if installed something few days ago many small softwares. I don't know if there was some virus or something but anti virus didn't tell anything. Now when i log ion then i keep getting many cmd screen opening and closing and i can't do anything. If i switch user and log with my admin acount then it also happens again witha dmin account. Now when again if i go back to old screen by switching user. then annoying cmd windows are not there. i want to know how can i fix that I managed to get print screen of that

    Read the article

  • Why suddenly DOS-type hexadecimal file names?

    - by Marvin Nicholson
    One of the fairly recent folders on my XP SATA data drive suddenly shows DOS-type hexadecimal file names (i.e., eight characters with three-character extensions) I deleted them and now my Recycle bin shows them with a tilde (i.e., 194ABE~1.JPG). The images are all valid but the file names I assigned are gone. (The 2-terabyte SATA data drive has no OS, if that matters.) The last time this happened on an IDE drive, I was able to back up all the remaining files just before the drive died. Am I facing the same scenario now with my 2-terabyte SATA data drive? It is only a couple of years old. Should I quickly buy another one and back up 20 years of files to it before my current drive dies?

    Read the article

< Previous Page | 381 382 383 384 385 386 387 388 389 390 391 392  | Next Page >