Search Results

Search found 9620 results on 385 pages for 'backup profile'.

Page 253/385 | < Previous Page | 249 250 251 252 253 254 255 256 257 258 259 260  | Next Page >

  • Iptables mark incoming packet - vpn routing

    - by Tom
    I have connected my home to my workplace for out of house backup reasons through openvpn. The connection is working nicely. At work I have 5 fixed IP addresses. Now I would like to assign one of these IP addresses to be forwarded to my home machine. I have confirmed packet arrival at my home machine with tcpdump. The problem is that my default route at home is NOT the tun0 (naturally), but eth0 to my own ISP. So I created a separate routing table to route my tun0 packets back to where they belong, but do not how to mark the incoming packet which arrive through tun0 with iptables, so I can drive them back. I do not want any port restrictions, but only what comes from tun0 should leave through tun0 thanks tom

    Read the article

  • Why do all my browsers default to the "Impact" font?

    - by J Alain Ferry
    I've been struggling with a font issue which I hope you can help resolve. A few weeks ago I started seeing "Impact" as the default font on lots of webpages, rendering much of the content illegible. I think this issue began after I installed other fonts, but I'm not sure. The issue occurs on Chrome, IE and Firefox which suggests it is a Windows issue. By the time I tried to revert the changes via System Restore, it was too late. I tried to reset the default font but Windows does not seem to respond to that. When I click "Restore to default font settings" the button greys out but nothing else happens. No dialog box and no change in the issue. I've tried deleting the fonts in my font folder and re-installing them (from backup) but that did not work either. I inititially documented the issue here: http://forums.techguy.org/windows-7/1020796-help-weird-font-all-browsers.html Any ideas? Thanks!

    Read the article

  • SSH not working after Restoring Running-Config to a Replacement Cisco Router

    - by Kyle Brandt
    One of my Cisco routers died over the weekend, Cisco sent the replacement and I restored the the config using copy tftp: running-config. Everything seems to work fine but I can no longer ssh into the router (I can telnet). The connection is refused, so it isn't listening on port 22 it seems like. I had previously backed up the config by just doing ssh router 'show run' > backup_config from my workstation. So: Is there anything wrong with my method of backup vs copy running-config tftp:? I know I haven't given any debug information, but is there something typical I need to do to get ssh working?

    Read the article

  • Demoted domain controller and now IIS has permissions issues

    - by tladuke
    I have a machine that was a domain controller and everything else, includin an IIS site. I built new 2 new domain controllers and transferred the FSMO roles and waited a day and then demoted the original domain controller. Now the IIS site says : HTTP Error 401.2 - Unauthorized You are not authorized to view this page due to invalid authentication headers. I have a call in with the web app vendor, but maybe someone can guess what I need to fix now. I haven't looked at IIS since this was installed and am pretty lost. I thought about restoring the machine from a backup, but I think that would be an Active Directory disaster, right? The server is Windows 2008 (not R2). The new DCs are 2008 R2

    Read the article

  • VMware Server guest systems are extremely slow with IO load on host (Ubuntu 8.04)

    - by Dennis G.
    We are experiencing performance issues with a VMware Server 2.x installation on an Ubuntu 8.04 host. When the host system is generating IO load (for example, copying large files as part of a backup operation), the guests (also Ubuntu 8.04) become extremely unresponsive and slow (simple Apache HTTP requests taking 5 seconds instead of the usual 200ms). We tried optimizing various aspects of the VMs, but the issue remains. Is there a known bug with VMware performance under linux if host IO load is high? Is there a way to fix this? Is this only an issue with Ubuntu systems, or have you seen it on other systems before? Thanks!

    Read the article

  • Php template caching design

    - by Thomas
    Hello to all, I want to include caching in my app design. Caching templates for starters. The design I have used so far is very modular. I have created an ORM implementation for all my tables and each table is represented by the corresponding class. All the requests are handled by one controller which routes them to the appropriate webmethod functions. I am using a template class for handling UI parts. What I have in mind for caching includes the implementation of a separate Cache class for handling caching with the flexibility to either store in files, apc or memcache. Right now I am testing with file caching. Some thoughts Should I include the logic of checking for cached versions in the Template class or in the webmethods which handle the incoming requests and which eventually call the Template class. In the first case, things are pretty simple as I will not have to change anything more than pass the template class an extra argument (whether to load from cache or not). In the second case however, I am thinking of checking for a cached version immediately in the webmethod and if found return it. This will save all the processing done until the logic reaches the template (first case senario). Both senarios however, rely on an accurate mechanism of invalidating caches, which brings as to Invalidating caches As I see it (and you can add your input freely) a template cached file, becomes invalidate if: a. the expiration set, is reached. b. the template file itself is updated (ie by the developer when adding a new line) c. the webmethod that handles the request changes (ie the developer adds/deletes something in the code) d. content coming from the db and ending in the template file is modified I am thinking of storing a json encoded array inside the cached file. The first value will be the expiration timestamp of the cache. The second value will be the modification time of the php file with the code handling the request (to cope with option c above) The third will be the content itself The validation process I am considering, according to the above senarios, is: a. If the expiration of the cached file (stored in the array) is reached, delete the cache file b. if the cached file's mod time is smaller than the template's skeleton file mod time, delete the cached file c. if the mod time of the php file is greated than the one stored in the cache, delete the cached file. d. This is tricky. In the ORM implementation I ahve added event handlers (which fire when adding, updating, deleting objects). I could delete the cache file every time an object thatprovides content to the template, is modified. The problem is how to keep track which cached files correpond to each schema object. Take this example, a user has his shortprofile page and a full profile page (2 templates) These templates can be cached. Now, every time the user modifies his profile, the event handler would need to know which templates or cached files correspond to the User, so that these files can be deleted. I could store them in the db but I am looking for a beter approach

    Read the article

  • Jungledisk - choosing between Amazon S3 and Rackspace ?

    - by Dietmar
    Hi, jungledisk is owned by rackspace and offers an option to choose Amazon S3 which charges for traffic whereby if data is stored on rackspace, no charges for traffic. For a small company, to just backup files with jungle disk from ubuntu linux OS computers, i wonder what to choose. Commercial psychologically the price on traffic and the Amazon brand make it appear or maybe it is, of better quality than the cheaper Rackspace offer which is when it comes for just backing up ok ? but if its about backing up around 30 GB which updates from time to time, what would you consider a good choice ?

    Read the article

  • Outlook items suddenly disapears from inbox after synchronization

    - by Jan Šotola
    We had some disk problems on Exchange server. We had to fix the Exchange 2007 using eseutil. Everything works fine excpet of following problem with one e-mail account: All inbox messages of mentioned e-mail account are fully accessible through the OWA. When any user creates new Outlook profile (using .OST mode) and starts synchronizing, all inbox messages are visible in the Outlook's inbox folder untill the Inbox synchronizing is completed. Once the synchronizing progress bar reaches 100% (and the Outlook starts synchronizing Sent items or another folder), most of messages disappears from the inbox view of the Outlook. The .OST file size does not reduce. New mails (delivered after Exchange server crash) remains all in the inbox. There is no special view or filter applied in the Outlook. The problem has been reproduced on two different computers (Vista+Office2007, Win7+Office2010). We assume there is some problem on the Exchange server but we don't know how to fix it. Can anybody give us some advice?

    Read the article

  • "Bad response to Storage command" when scheduling job with Bacula

    - by Joril
    I have a Bacula setup with 9 clients, and it's working happily. Today I had to add another client, so I went and copied+adapted the existing configuration files from another client, but when I schedule a job for the new client, I get these errors: 20-Mar 17:50 tools-dir JobId 39: Start Backup JobId 39, Job=BackupPresenze2.2012-03-20_17.50.49_04 20-Mar 17:50 tools-dir JobId 39: Using Device "FileStorage" 20-Mar 17:50 presenze2-fd JobId 39: Fatal error: Failed to connect to Storage daemon: bacula.mylan.local:9103 20-Mar 17:50 tools-dir JobId 39: Fatal error: Bad response to Storage command: wanted 2000 OK storage , got 2902 Bad storage From the client I can telnet to bacula.mylan.local:9103 just fine, and jobs for other clients work successfully... What could I check? (Server and client run Ubuntu 10.04, if it's relevant)

    Read the article

  • Link bonding across multiple switches?

    - by Bryan Agee
    I've read up a little bit on bonding nics with ifenslave; what I'm having trouble understanding is whether there is special configuration needed in order to split the bonds across two switches. For example, if I have several servers that all have two nics each, and two separate switches, do I just configure the bonds and plug 1 nic from each into switch #1 and the other from each into switch #2? or is there more to it than that? If the bonds are active-backup, will a nic failure on single machine mean that server may become disconnected since the rest of the machines are using the primary nic and it's using the secondary? Or do you link the switches with one cable as well?

    Read the article

  • Vmstat indicates memory is disappearing

    - by jimbotron
    I wanted to profile the memory usage of a script. Here's the output before it was running: procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 0 0 15624 186660 39460 439052 0 0 0 2 1 1 0 0 100 0 Here's the output while the script is running, at the point where free memory was at its lowest value: procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 0 0 15624 11464 40312 473524 0 0 0 2 1 1 0 0 100 0 So free memory dropped by about 175 MB, and I expected that buff would increase by that amount. But it seems the other columns changed by relatively negligible amounts - how is this possible? Am I interpreting this wrong, or is some memory just not being accounted for in this output?

    Read the article

  • Firefox 3.6 performance increase tricks.......

    - by metal gear solid
    I use many Add-ons which helps me in Web development so i can't uninstall those addons. and usually I keep open lots of tabs in Firefox. And almost always keep Firefox on on my system. I use default profile only I always keep every addons and firefox itself updated. I found some addons to reduce memory use on Firefox addons site , but user reviews were not good for them still Is there any tested tricks to increase performance and reduce memory use of Firefox 3.6, which really works?

    Read the article

  • Windows 7, Drivers, Cloning and Sysprep

    - by Frank Thornton
    I have Windows 7 working on one machine but when I move it off to the new machine it crashes. If this is a driver issue do I need to download the drivers and install them first? Or do I install Windows 7 on the new machine first then copy the drivers folder over to the other disk so it will boot up correctly? EDIT: I was trying this as well: http://www.todo-backup.com/support/tutorial/redeploy-system-to-dissimilar-hardware.htm EDIT: I tried sysprep but on bootup I see the Windows logo then the system crashes. I can stick the drive back in the old hardware and it runs fine.

    Read the article

  • Massive amount of subfolders and long subfolders. ¿How can I delete all of them?

    - by Carlos
    Good day. We have a little problem here. We have a share with the backup of all the server's offices, Its a really big share with more than 8.000.000 files. Our users usually give long names to the folders they create, and then make subfolders (long too) and more subfolders... and more suboflders.... We have a new share with more capacity, and with a simpe robocopy bat we copied all the files and folders (some give problems, but we manually copied them) But the problem is deleting them. del command didnt work well when so long paths, neirder rmdir... I'm tried some commanders, but no luck. Can u recommend me any tool that can delete recursively or able to delete 255+ paths? Edited: The SO on background of the share it's NetApp OS. But I can access it from Windows Servers. 2000 and 2003 Thanks.

    Read the article

  • Outlook - Out Of Office can't set

    - by derjur
    Exchange 2007, Outlook 2010 Error when setting Out Of Office (Automatic Replies) from Outlook: Your automatic reply settings cannot be displayed because the server is currently unavailable. Try again later. Setting this through OWA works fine. Autodiscover is working, and all addresses are resolvable through DNS. Outlook profile created from scratch. Anonymous access disabled for EWS/OWA (IIS on Exchange box). I'm out of steam for troubleshooting. Any ideas?

    Read the article

  • Can I import contacts from an Exchange 2003 EDB file for a single user?

    - by Drarok
    We have recently had to reinstall a very unhappy Exchange 2003 server for a client, and whilst rebuilding their server we moved them onto a temporary one. During the course of all this, ExMerge was used due to Windows being so broken that none of the backup software could run. As a number of the user mailboxes were way over 2GB, we had to do date ranges of a few months at a time to avoid ExMerge's 2GB limit. We went back as far as 2006, as the files produced by ExMerge seemed empty that far back. Unfortunately, one of the users has reported that around 2 3rds of their contacts are missing, and that pre-2006 sounds about right for the missing items. Is there any way I can mount the old EDB file into Exchange, or otherwise read their contacts into a usable format? The server is running Windows Server 2003 SBS R2 (SP2), and Exchance 2003 (SP2, I think).

    Read the article

  • sql server log shipping

    - by voam
    I am setting up log shipping with two 2008 sql servers connecting with a vpn. As far as I know as long as the sql agent is able to access the share on the primary/secondary servers everything will work. When I set up the Log shipping on the primary server using SQL Mangagement Studio, before I can set any of the "Secondary Database Settings" it asks me to connect to the secondary server. But I really don't want to open up the connection to the secondary server. I will be initialing this secondary database with a backup so as long as the transaction logs get copies everything should work. How to I work around the GUI not enabling any of the settings for the secondary server until I actually connect to it? Thanks in advance!

    Read the article

  • What does advantages of WIM image that file based vs. ISO or another sector based images in Administrator

    - by saber tabatabaee yazdi
    answers of this question is very important to administrators to choose between Microsoft solution or other 3rd Party Tools. WIM solution is file based and others are sector by sector images like Acronis or other backup and clone tools like Norton Ghost or Cloning in vmware ESXi and etc. My question about the advantages of WIM to edit or deploy in Enterprise? to get agreements of managers to choose between these alternative decisions. Are we had any Key Performance Indicators in Change management? Are there any statistics about compares of these solutions? What Time is spend in any solotion to get image or change images or revert and apply images? Is it possible to change and modify images that made in All of these solutions as easy as in WIM file?

    Read the article

  • Where to export a truly global environment variable?

    - by Socio
    I want to set an environment variable that will be visible to all processes launched by Upstart. This is on a CentOS system, but I presume the same applies to Ubuntu given that they both use Upstart. Somewhere in /etc/init/ perhaps? Note that adding it in /etc/profile.d only applies to login shells. I want all processes (e.g cron, rc.local, etc) to see it. Obviously I'd prefer not to edit existing sys config files if it can be avoided.

    Read the article

  • Any way to program a system which automatically restores home / sql database

    - by Mirage
    I have made two shell scripts Script 1: It does all Home directory backups with name username_home_date.tar.gz Script 2: It does SQL backups of all sites every 3 hrs. username_databse_date.sql.gz Now currently if I want to restore the site, I have to copy the tar file to /home/username and then untar there with all the permissions as well and then manually import the database. Is there any way (for instance a program, system or script) that I can just select which backup I want to restore and do automatically? Maybe like a cPanel addon thing.

    Read the article

  • Windows Mobile deletes my Contacts when deleting a partnership

    - by bitbonk
    Sometimes when I reinstall my PC get a new Work PC or buy a new home PC ,ActiveSync, now the Device Center asks me to delete an existing partnership to setup a partnership with the new PC, since only two partnerships are allowed. When I delete that partnership all contacts that originally came from that partnership get deleted too. How can I prevent that from happening. Do I really always have to remember to frequently backup all my contacts in a safe place? How much redundancy is needed. I hat them on one of my PCs and on my phone.

    Read the article

  • Merely chainloading an Acer Recovery Partition deleted all data

    - by WindowsEscapist
    I was starting a backup of Acer's factory restore partition located inside of an extended partition to determine whether or not it still worked. I clicked "take no action" once I saw that it had, in fact, successfully started up. However, when I rebooted, I got an "error: no such partition" and was dropped to a GRUB recovery prompt. Upon further investigation, I discovered that all partitions inside the extended partition were gone except for the recovery partition! What happened? How can I fix this? testdisk doesn't find the deleted partitions!

    Read the article

  • Programs keep waiting for external disk to spin up - how to ignore disk?

    - by Andrew J. Brehm
    Like many Mac users I have an external Firewire disk hooked up to my Mac to be used by Time Machine. This works very well, backup-wise. The problem is that very often when I use a Mac application and try to open a file, the file selection dialogue window hangs until the external disk has spun up. I never ever want to open a file on the external disk. Sometimes this happens even when I just want to save a file I already saved (i.e. type something and press meta-s). Is there anything I can do about this?

    Read the article

  • File corruption when copying different file on raid 1

    - by Stephan
    I have a RAID 1 configuration of 2 1TB drives on a Fedora 12 box. Most of what is stored there are video files that are numerical labeled. The problem I'm having is that I had one of the video files get corrupted. I copied a replacement from a backup and replaced the bad file and now it works fine. However, after doing this the next numbered file goes from 350MB to 200KB and all but about .5 second of video disappears. If I then replace that file it happens to the next one down the line. Ex: Replace corrupt file 1.avi and file 2.avi shrinks to 200KB. Replace now corrupted 2.avi and it works but 3.avi gets screwed up. I have run SMART tests on the drives and they report fine. Does anyone have any tests I can run to try to figure out what is going on? EDIT: It is a two disk software RAID 1 with an ext4 filesystem

    Read the article

  • Generating my own SQL Server SSL Certificate

    - by Haoest
    I hear it's possible to make myself a test certificate with MakeCert from Windows SDK, but the beast is 1.5 GB, which I feel reluctant to download for a half-megabyte program. I do, however, have IIS Resource kit with me, which has selfssl to generate certificate for IIS. Will that work for SQL Server? Is there a more convenient way of generating myself an SSL certificate for SQL Server use? I even tried using CREATE CERTIFICATE with TSQL within SQL Server and then have it BACKUP into a file, but with no success. I must have confused the concept of certificate fundamentally. Any advice?

    Read the article

< Previous Page | 249 250 251 252 253 254 255 256 257 258 259 260  | Next Page >