Search Results

Search found 12104 results on 485 pages for 'auth basic'.

Page 228/485 | < Previous Page | 224 225 226 227 228 229 230 231 232 233 234 235  | Next Page >

  • Memeory Leak in Windows Page file when calling a shell command

    - by Arno
    I have an issue on our Windows 2003 x64 Build Server when invoking shell commands from a script. Each call causes a "memory leak" in the page file so it grows quite rapidly until it reaches the maximum and the machine stops working. I can reproduce the problem very nicely by running a perl script like for ($count=1; $count<5000; $count++) { system "echo huhu"; } It is independent of he scripting language as the same happens with lua: for i=1,5000 do os.execute("echo huhu") end I found somebody describing the same issue with php at http://www.issociate.de/board/post/454835/Memory_leak_occurs_when_exec%28%29_function_is_used_on_Windows_platform.html His solution: Firewall/Virus Scanner does not apply, neither are running on the machine. We can also reproduce the issue on other Developer Machines running XP 64, but not on XP 32 Bit. The guilty guy for the allocation is C:\WINDOWS\System32\svchost.exe -k netsvcs which runs all the basic Windows services. Does anybody know the issue and how to resolve it ?

    Read the article

  • Redhat 6 gui installation VS kickstart gives me different packages?

    - by jonaz
    If i do the graphical install and select basic server + aide and screen i get a system with 535 installed packages. If i look at the /root/anaconda-ks.cfg file in that freshly installed system i see: %packages @base @console-internet @core @debugging @directory-client @hardware-monitoring @java-platform @large-systems @network-file-system-client @performance @perl-runtime @security-tools @server-platform @server-policy @system-admin-tools pax python-dmidecode oddjob sgpio certmonger pam_krb5 krb5-workstation nscd pam_ldap nss-pam-ldapd perl-DBD-SQLite aide screen If i then install a NEW system using a kickstart only containing those packages i get 620 installed packages. So basicly my question is why does the system install almost 100 more packages when using kickstart compared to the GUI installation when the exact same packagegroups are selected?

    Read the article

  • force https with apache before .htpasswd

    - by johnlai2004
    I have this in my .htaccess file RewriteEngine On RewriteCond %{HTTPS} off RewriteRule ^(.*)$ https://www.myweb.com/phpmyadmin$1 [R,L] AuthUserFile /var/www/myweb/.htpasswd AuthGroupFile /dev/null AuthName "Sovereign Databases" AuthType Basic <Limit GET> require valid-user </Limit> But everytime I go to http://www.myweb.com/phpmyadmin, the .htpasswd prompts me for a credentials BEFORE i'm redirected to https://www.myweb.com/phpmyadmin. After I type in my username and password, I get redirected to https://www.myweb.com/phpmyadmin. The problem is that I don't want anyone to submit their username and password unencrypted via http. How do I force people to login via the https version even if they typed in the http version?

    Read the article

  • Suggestions for wireless adapter

    - by Michael Kniskern
    I recently purchased a desktop computer and it did not come with a wireless network card. I am currently using the Belkin Wireless G USB Adapter and it is a very sub par product. It has very slow response times and download speeds with just basic browsing and downloads podcasts through iTunes. Does anyone have a better suggestion for a wireless adapter? Should I go with another USB adapter or one that connects directly into the motherboard? I am current using Windows Vista 64 bit Home Premium.

    Read the article

  • Out Of Memory Error - Magento

    - by robobobobo
    Ok normally I understand when my server is giving me out of memory errors, but this one has me stumped! I'm running a magento based site, with one or two plugins in it and the rest is pretty basic. The site runs and loads fine wiht no issues. However in the backend - Configuration - Payment Methods it gives me the following out of memory error Fatal error: Out of memory (allocated 39059456) (tried to allocate 85 bytes) in ########/Varien/Simplexml/Element.php on line 84 Now this is where I'm confused..it's allocated more than it tried to allocate? Am I correct there? So how is it running out of memory? My server has 6Gb ram, an SSD and 2 CPU's running WHM with a few other low traffic sites on it. I set my php memory limit to 100mb, 1000mb and finally unlimited but all to no avail! I'm completely lost here, would really appreciate some expertise on this Cheers

    Read the article

  • Which iptables rule do you think is a 'must have'

    - by Saif Bechan
    I have some basic iptable rules set up now for my vps. Just block everything except some default ports, 80,21,22,443. I do get brute forced a lot. I have heard that iptables is very powerful but I have not seen many use cases. Can you give me an example of a(some) rule(s) you always use and give a small example why. I can not find a general best practice post here on SF, if there is any I would like the link. If this is a duplicate I am sorry and it can be closed.

    Read the article

  • VPN within a VM to allow for internet access on the host

    - by David Durrant
    I have a network connection (created under Networks and Sharing) that I use to connect to a customer's site. But when I use this to connect to the site, I loose all access to the public internet, and can only access customer specific items. I want to circumvent this issue by creating a VM and then utilizing the VM to connect to the network location and interact within the customer's domain, while leaving my host machine open to the internet. I'm not extremely familiar with networking, but I have a few basic skills. Please let me know if this is possible and what the correct procedures are. I already have a VM created with VirtualBox, and both the host and guest are running Windows 7 x64. I have created duplicate VPNs already, but can only connect successfully on the host machine.

    Read the article

  • Copy a file from source directory to target base directory and maintain source path

    - by Citizen Dos
    Forgive me, I am probably not using the right terms to describe the problem and misunderstanding the most basic usage for a couple of common commands. I have a simple find statement that is locating files that I want to copy. I want to tack on the -exec cp {} and have cp copy the file from the source directory to a new base directory, but include the full path. For example: "find . -name *.txt" locates /user/username/projects/source.txt "cp {} [now what?]" copies the file to /user/newuser/projects/source.txt

    Read the article

  • How does the data storage work? [closed]

    - by Andres Adhi
    I am really new to the whole concept of Data storage, Domain, Server and everything else related to this. Can someone pleases explain what a Domain is? How are server part of the Domain and How are Database stored in the Server or Domain? How does a new server be able to connect to existing database server to get all the data needed. I tried to find this information in the web but I am not really finding a good resource. It may be because these is really basic information. I will really appreciate if someone can explain these concept in plain terms. Thanks in advance.

    Read the article

  • windows server backup 2008 R2 - what is generating all the change data?

    - by bobjandal
    We have a small relatively idle windows server 2008 R2 installation that does basic filesharing and exchange for about 10 not very active users. When running a windows server backup, the incremental data daily is about 20GB. This is not coming from users shared files, nor from changes in their mailbox sizes. The total size of the installation is 249GB, which is mostly old files. Where is all this data coming from, and how can I reduce it ? Using online backup of the vhd file from the backup is taking a while because of this daily change. Is there some way I can at least see what files are changing and contributing to this data ? Options I can think of but am not sure about: 1) pagefile churning - altho the backup does not include the pagefile, perhaps the changed blocks left behind are included ? 2) logs or something ? but the installation size stays the same every day 3) should I zero free space using sdelete before backing up perhaps ?

    Read the article

  • Volume size doesn't match Disk size after gparted expansion

    - by Cybersylum
    I just expanded a basic disk on a Windows XP VM from 15gb to 40 gb using GPARTED LiveCD (0.5.2-11). I didn't notice anything unusual during the expansion; but after I rebooted back into Windows, the disk capacity doesn't match the disk size as it should (only 1 volume on the disk). The disk shows as 40gb; but the C: volume still shows the original size. I've tried expanding the disk again with GPARTED (no change), and using VMware converter and have it adjust the size of the volume during the process (complains about a lack of space of snapshot error inside the os). The volume has 27% free space so I don't think it is a space issue. Chkdsk doesn't seem to find anything wrong either. The OS seems to run just fine, it doesn't see the additional space however. Any ideas?

    Read the article

  • media is write protected when using diskshadow.exe, start-bitstransfer powershell cmdlet

    - by Aaron - Solution Evangelist
    i am trying to use the powershell start-bitstransfer cmdlets to transfer a file i have exposed using a vss snapshot (via diskshadow), but unfortunately i am receiving the following error: Start-BitsTransfer : The media is write protected. At line:1 char:49 + Import-CSV c:\hda1\bits.txt | start-bitstransfer <<<< -transfertype upload -Authentication "Basic" -Credential $cred + CategoryInfo : InvalidOperation: (:) [Start-BitsTransfer], Exception + FullyQualifiedErrorId : StartBitsTransferCOMException,Microsoft.BackgroundIntelligentTransfer.Management.NewBits TransferCommand we really want to utilize the bits endpoint we are attempting to transfer the files to. is there any other way we can go about this (aside from copying the files elsewhere first, unless we can copy one slice at a time and transfer that)?

    Read the article

  • Shell Script, iterating over a folder

    - by Martin
    I have very basic shell scripting knowledge. I have photos under original folder on many different folder like this folder + folder1 + original + folder2 + original + folder3 + original + folder4 + original Using mogrify I'm trying to create thumbs under a thumb folder following a structure to this. folder + folder1 + original + thumb + folder2 + original + thumb + folder3 + original + thumb + folder4 + original + thumb I'm a little lost in how to write the shell script that may iterate through it. I'm ok giving mogrify its settings but I don't complete understand how to tell the script to go iterate each folder to run the mogrify command.

    Read the article

  • Tunneling HTTP traffic from a particular host/port

    - by knoopx
    Hello, I'm trying to figure out how to access from my development machine (Devel) to a third party web service (www.domain.com) which I am not allowed to directly contact using my office IP address. Here's a basic diagram (i'm not allowed to post images...): http://yuml.me/diagram/scruffy/class/%5BDevel%5D-%5BA%5D,%20%5BA%5D-%5BB%5D,%20%5BB%5D-%5Bwww.domain.com%5D The only machine allowed to access that service is B (production server) but I do neither can directly access it from my development machine (Devel). So in order to access the web service I have to ssh into A, and then from A to B to access www.domain.com Is there any way of tunneling traffic from B to A and then back to my development machine so I can directly access www.domain.com without having to ssh into every box? Devel: My development machine. A, B: Linux servers. I own root access on both. B: Production server www.domain.com: Third party HTTP API production server uses.

    Read the article

  • How to scan and annotate documents and receipts so metadata is searchable and embeddeded in document

    - by Precipitous
    I'm getting tired of living in the modern age and still having a messy filing cabinet full of paper. I have a nice cheap scanner and want to scan receipts, warrenties, and so forth. To make it possible to find these, I'd like to make them searchable via Windows Search (or google desktop search). Because I'm lazy with backups, I want a file format that embeds metadata into the file. Backups should be as simple as coping a bunch of image files around. I'm also cheap, and am not going to install an expensive solution. I'm hoping one of the basic formats can do this. So, what common scannable format satisfies: searchable and and metadata embedded in image generated. The more "open" and portable the better.

    Read the article

  • Audio and video streaming using Network Simulator in Linux

    - by Parth_90
    I am working on a project which is to show the simulation of streaming of audio and video data in wireless networks. I want to show the simulation that involves a base station, with few wireless stations. The base station should start sending data once it computes a certain value . On receiving the data, each wireless must begin communicating with the base station. I have gone through basic NS-2 tutorials from over here but I am not getting how to go about integrating it with my project. Can anyone tell me how to do it using NS2 or any other network simulator?

    Read the article

  • Spanned volumes on new install

    - by Noio
    My Windows 7 Release Candidate is about to expire, so I'm going to do a clean install of a retail version. I have two volumes, on four physical drives, as follows: Disk 0: Spanned Volume (D:) Disk 1: Primary Partition, Boot/Windows Install (C:) Disk 2: Spanned Volume (D:) Disk 3: Spanned Volume (D:) If I install Windows to a formatted drive 1, will it still recognize the spanned volume in Disks 0, 2, and 3? The spanned volume is not redundant in any way, so the volume is 1.5TB consisting of three 500GB disks. I don't have the space to do an external backup, and I thought it was impossible to convert a spanned volume back to a basic volume.

    Read the article

  • Dell Studio XPS 16 Runs Hot

    - by dtbarne
    Specs: 1920x1080 i7 1.6 GHz quad core 6GB RAM 1GB ATI Radeon HD 6570M/5700 Series 500GB 7200rpm Hard Drive Love this laptop for many reasons, but it constantly runs hot (CPU is low 70s with basic tasks and often 80+ is not uncommon) and I'm finding it too much to deal with. The laptop feels very hot (almost too hot for a lap) and often gets so hot that the OS slows down or freezes altogether. I've tried cleaning it out and even replacing the thermal paste. I often use an external cooler, but it only helps 3-5 degrees and it's a pain to have to use. I've come to the conclusion that it just runs hot. I have two questions: What is to blame? The i7 processor, the gfx card, or is it just that this laptop has poor cooling? Does the Dell XPS 15 run cooler? I'm looking at replacing my current laptop, but I don't want to run into the same problem.

    Read the article

  • subversion: enforce TLS

    - by Daniel Marschall
    Hello, I am running subversion on a Debian Squeeze system with Apache2 and mod_dav for viewing the contents with a webbrowser. I want to enforce the usage of TLS, so that the login data and the SVN contents cannot be read from the connection. I have tried following: <Location /svn> DAV svn SVNParentPath /daten/subversion/ # our access control policy AuthzSVNAccessFile /daten/subversion/access_control # try anonymous access first, resort to real # authentication if necessary. Satisfy Any Require valid-user # how to authenticate a user AuthType Basic AuthName "Subversion repository" AuthUserFile /daten/subversion/.htpasswd # Test SSLRequireSSL RewriteEngine On RewriteCond %{SERVER_PORT} !443 RewriteRule ^svn/(.)$ https://www.viathinksoft.de/svn/$1 [R,L] </Location> at file /etc/apache2/conf.d/subversion.conf Alas, this does not work. There is no redirect and there is still a HTTP request working at /svn/(projectname)/(somefolder) . This SSL-enforce-policy should work for - viewing the contents with webbrowser - retrieve contents with TurtoiseSVN client - committing contents with TurtoiseSVN client Can you please help me? Regards Daniel Marschall

    Read the article

  • Windows Server 2008: Limit UDP/TCP packets per IP or ban

    - by WBAR
    How I can limit UDP/TCP packets per IP send to my host (or better PORT) per second or minute ? Would be nice to ban that IP for 12/24 hours or even for ever. I got Windows Server 2008 and I'm very poor in Windows administration but quite good in Linux. EDIT: By basic problem is that They sending a lot of rubbish UPD and TCP packets.. TCP packets without SYNCH, fragmented UDP packets so my servers stop responding.. So I need to cut off users (IPs) sending more than X packets per second. I need solution witch provides me, somehow, configurable: X packets of certain type (UDP, TCP or both - lets say parameter named Z ) are allowed to be received by IP on Y port, otherwise this packet should be DROPPED. My virtual hosts are hosted by VirtualBox and I'm able to forward all incoming packets certain type and certain port to the specific Virtual Host, but I need to DROP them before my VirtualBox receive them.

    Read the article

  • Mysql Master-ColdMaster

    - by enedebe
    I explain my case: I'm at Amazon AWS and I want to be fault tolerant on a entire region failure. My basic problem is to have the db in sync with 2 regions. My options: Master-Master (high lag) Hand made sync every 5 minutes Master-ColdMaster?! (copy on the fly but Master won't wait the other region commit) In my system we could afford loosing a piece of data (we're not a bank) the last inserts in the db, but we could not afford more than 10 minutes of downtime. The database is small and the level of inserts is low, and I wouldn't affect the normal usage waiting other region commit. Is the 3 solution posible? And the most important, once the primary fail how we can detect and change the rol between master-coldmaster -- coldmaster-master ? Is there any clean-mode to restore between failure? Thank's!

    Read the article

  • MS-DOS 7.1, Active-Boot

    - by user3005255
    I created a MS-DOS 7.1 (Win 98) active-boot USB-flash stick and it functions properly. Nevertheless I still have some unsolved issues: I want to change my keyboard-layout to german but I don´t know where to start since I only have the "basic-DOS" with IO.sys, MSDOS.sys + command.com. I already created Autoexec.bat + config.sys but don´t know which files I need and which I have to change in order to get a german keyboard. Also I can´t really shutdown my DOS properly, what is the right way to do it in MS-DOS 7.1? Thanks very much in advance for your help!

    Read the article

  • Amavis / Atermime - Dynamic Email Signatures (Disclaimers)

    - by Frank
    I am running a basic iredmail install with amavis, postfix, dovecot. I want to use the declaimer feature to automatically append a disclaimer (in our case a signature). The reasoning for this is that clients (ipad, outlook, etc) don’t have to be configured with signature settings, and that we have corporate control over the actual signatures that are posted. Instead of: '[email protected]' => 'boss.domain.com.txt' I am trying to get to %mailbox + ‘@’ + %domain => %mailbox + “.” + %domain + ‘.txt’ This way we can simply use a script to generate the .txt / .htm signature files using just mysql / php. THANKS IN ADVANCE

    Read the article

  • Web Folder size/quota reporting tool?

    - by nctrnl
    I am currently using a Visual Basic script to determine how big the web folders are and what quota is decided for each folder. The quota is in no way a physical limit, just a value inserted by me to decide whether a user is using too much space or not. The script does the job quite neatly and sends an html file by mail on a regular basis. The problem is that it's such a hassle to insert new quotas since I have to fiddle around with the code. A central "control panel" with an overview and ability to insert new quotas would be more suitable. Is there any software that can do the following: Scan specified folder/subfolders Report the file size and present it in some sort of interface (could be a php/mysql solution) Ability to specify a quota and see the difference value ? It is really important that the quota handling is made simple so that some non-technician can handle this.

    Read the article

  • Web Folder size/quota reporting tool?

    - by nctrnl
    I am currently using a Visual Basic script to determine how big the web folders are and what quota is decided for each folder. The quota is in no way a physical limit, just a value inserted by me to decide whether a user is using too much space or not. The script does the job quite neatly and sends an html file by mail on a regular basis. The problem is that it's such a hassle to insert new quotas since I have to fiddle around with the code. A central "control panel" with an overview and ability to insert new quotas would be more suitable. Is there any software that can do the following: Scan specified folder/subfolders Report the file size and present it in some sort of interface (could be a php/mysql solution) Ability to specify a quota and see the difference value ? It is really important that the quota handling is made simple so that some non-technician can handle this.

    Read the article

< Previous Page | 224 225 226 227 228 229 230 231 232 233 234 235  | Next Page >