Search Results

Search found 22900 results on 916 pages for 'pascal case'.

Page 420/916 | < Previous Page | 416 417 418 419 420 421 422 423 424 425 426 427  | Next Page >

  • Why is my ethernet interface in promiscuous mode

    - by nhed
    I read that seeing a flag of M in netstat -i is the way to tell which of your interfaces is in promiscuous mode I run it and I see that eth1 is in promiscuous mode $ netstat -i Kernel Interface table Iface MTU Met RX-OK RX-ERR RX-DRP RX-OVR TX-OK TX-ERR TX-DRP TX-OVR Flg eth1 1500 0 1770161198 0 0 0 57446481 0 0 0 BMRU lo 16436 0 97501566 0 0 0 97501566 0 0 0 LRU This seems to be the case on all the machines I checked (All Centos6.0, both virtual and physical), any idea why ethernet devices would be in such a mode unless someone was running any pcap based app (sudo lsof | grep pcap shows nothing)? I did not see any mention of promiscuous in any of the config files (sudo grep -r promis /etc) Any ideas what puts the interface into that mode and why? p.s. most of the posts I see seem to be security related, this is not that

    Read the article

  • Corrupted attachments in Thunderbird forward filter

    - by Guandalino
    I created a filter in Thunderbird 14 on client A that, when a certain rule on incoming mail is satisfied, the same mail is forwarded to client B and C. Matching mails received from A are always forwarded to B and C, and that's what I want. Though, when the mail on A has an attachment, it is also forwarded by B and C but the file they receive is corrupted. In my case the incoming file on A is a Word file of about 10kb in size; clients B and C receive a Word file of 27 bytes. I think this could happen because the forward starts from A when the attachment is not yet completely downloaded. I'm not sure this is the cause, though. And I wouldn't have any idea for a fix. Any idea about how to investigate or solve the issue?

    Read the article

  • Suggestions on providing HA access to an external (fibre) RAID subsystem

    - by user145198
    We are looking at upgrading our storage capacity with an external RAID subsystem that has redundant (2) fibre controllers, each controller has 4 x 8 Gbps fibre ports. I would like to make access to this storage system occur via HA Linux. Ideally I would connect 2 fibre ports from each controller into each Linux server, and then export either NFS or iSCSI via a 10 Gbe interface. I have seen plenty of references to DRBD, however all of those references tend to use block storage that is solely attached to each machine, rather than having a shared block storage device, so I am unsure if DRBD could (or should) be used in this case. Ideas?

    Read the article

  • fail2ban on server with LXC Containers

    - by RoboTamer
    The issue is modprobe and iptables don't work inside an LXC Container. LXC is the userspace control package for Linux Containers, a lightweight virtual system mechanism sometimes described as “chroot on steroids”. iptables error inside the container is: # iptables -I INPUT -s 122.129.126.194 -j DROP > iptables v1.4.8: can't initialize iptables table `filter': Table does not exist (do you need to insmod?) Perhaps iptables or your kernel needs to be upgraded. I am guessing that it can't work because the LXC containers share one kernel, the main server kernel. How do I do fail2ban in this case. modprobe and iptables work in the main server so I could install it there and link to the logfiles somehow, my guess? Any suggestions?

    Read the article

  • Monitoring server room temperature

    - by user640
    I have a small server room with its own AC unit. Recently, the AC died, and the temperature increased from 70 F to 90 F. We rarely go in this room, so I was lucky that someone happened to notice that the fans were running a lot louder than normal as they walked past the door. It looks like I need a way to be notified when the temperature in that room gets too hot. What tools are you using to monitor the temperature in your server room? How does this tool notify you of a problem (email, SNMP, etc). Note: I've read this question on server temperature, but I'm interested in the whole room, not just the inside of a server case. Edit: Thanks for all the great responses so far! Many of these products measure much more than just temperature. What else should I be looking at and why?

    Read the article

  • Flow Fields in Dynamics NAV

    - by T
    I don’t know the exact business reason but someone asked how to identify all sales shipping orders with any negative quantities on them.  They needed to print these separately from the shipments that only have all positive quantities. Here is one way to solve this problem In NAV, open Sales Shipment Header Add a field of type Boolean (The field type should match the value you expect to return from the flow field method.  In this case we will use an exist so we expect a bool but we could have used a Sum, Average, Min, Max, Count, or a Lookup to return a value). Define it as a flow field and click on the ellipse next to CalcFormula Now [Has Negative Qty] is false if no negatives exist And [Has Negative Qty] and true if a negative exist So it is ready to be used as a filter on the report. Don’t forget that if you are using it in code, you may need to CalcFields. Hope that helps.  If you really can’t afford the field in your header, you can use code to check the lines for a negative value each time and use a skip or break function to skip that header record but it seems expensive to check them all if you only want a few to print. Please let me know if you think of a better solution.

    Read the article

  • Volume randomly turning itself down on Windows 7 64-bit

    - by Arda Xi
    This is the weirdest issue I've ever encountered with my PC. Every so often, my sound will start playing back at a lower volume. This happens when watching video, listening to music, all independently. It usually lasts anywhere up to a minute, after which it will turn up again. The weird thing about it is that the volume control in Windows remains at 100%, even though the volume is audibly a lot lower. (No, I'm not going deaf, it's just my PC. I checked.) I just have no idea where to go to troubleshoot this, even. I'm using Windows 7 64-bit with an on-board Realtek sound-card. Oh, just in case someone finds this question on Google or whatever, making sure this is on Do Nothing may fix your problem. Unfortunately, it did not work for me. My settings all seem fine. This is my audio slider. (Took the screenshot while the issue showed.)

    Read the article

  • sudo su - username while keeping ssh key forwarding

    - by Florian Schulze
    If I have a server A into which I can login with my ssh key and I have the ability to "sudo su - otheruser", I lose key forwarding, because the env variables are removed and the socket is only readable by my original user. Is there a way I can bridge the key forwarding through the "sudo su - otheruser", so I can do stuff on a server B with my forwarded key (git clone and rsync in my case)? The only way I can think of is adding my key to authorized_keys of otheruser and "ssh otheruser@localhost", but that's cumbersome to do for every user and server combination I may have.

    Read the article

  • Remap key combinations in Windows

    - by nickf
    This question is similar to this one, but the accepted solution doesn't work in my case. I work on two different laptops. On one, the Home and End keys are only available by pressing Fn+Left and Fn+Right. This can make it rather annoying when switching between the two since I always go to press Fn-Right and not get what I want. The tool suggested in the other question is Sharpkeys, however it doesn't recognise Fn+Right as a different key from Right, and I suspect that registry twiddling isn't going to help me here. (?) Is there any way to remap this combination? (I'm on Windows Vista if that helps)

    Read the article

  • Backing up data in an encrypted way

    - by Eli Bendersky
    I have the following use case: There's some data from my PC I want to periodically back-up online I own some hosting, so I want to use that for the backups, don't want to pay to another backup service I want to encrypt my data locally prior to moving it to the server I have no problem writing scripts to automate the process (say, periodically generate the backup and upload by FTP to my server), but my main question is about step 3 - the encryption: which way is recommended to encrypt my files (say, collected into a .ZIP) prior to uploading to the server? P.S. TrueCrypt seems popular but it's not quite what I'm looking for, since I don't want the files to be constantly encrypted here on my PC.

    Read the article

  • Cannot use second display with 12.04 and Intel 2000/3000

    - by Carolyn Marenger
    I am unable to get anything to display on my second monitor, or even get the system to recognize that there is a second screen. I am running Ubuntu 12.04 on a Gigabyte GA-H61M-S2PV, revision 2.0 box. The integrated chipset is an Intel 2000/3000, and there are both a D-Dub and a DVI-D display ports on the MB. This is the first operating system I have installed on this system. I have a second monitor plugged into the DVI-D port via a DVI-D to D-Sub adapter. I cannot verify that the motherboard or adapter were/are working, short of installing windows to test the theory. When I go into the "System Settings - Displays" control window, it shows one display. I have rebooted with the second monitor attached, and I have perused the BIOS settings in case it might have been disabled. So far, I have had no indication that the second monitor is recognized, not even a flicker at power on. If I swap monitors and cables between the DVI-D and D-Dub ports, the other monitor lights up, so I know the monitor and video cable are not the issue. Any suggestions would be appreciated. Thanks, Carolyn

    Read the article

  • Is it possible to have multiple subdomains point to the same Blogger blog?

    - by cclark
    For our application we want to have a status page which is hosted outside of the rest of our infrastructure so in case there are issues in our data center we can post updates for our users and our users will be able to access them. We registered a blog on Blogger and set it up with xyzstatus.blogspot.com and status.xyz.com. Everything seems to work fine. We need to perform some maintenance at our datacenter which will sever all connectivity so we're unable to have a redirect using nginx or apache. We'd like to do this with a short TTL CNAME DNS entry. Ideally www.xyz.com and app.xyz.com could be CNAMEd to status.xyz.com. When I setup the CNAME and go to that URL I get a Google broken robot 404 page. I figure I must need to let Google know it should associate traffic to www.xyz.com and app.xyz.com to the blog served up by status.xyz.com. But I can't see anywhere to do this in Blogger. Does anyone know if this is possible?

    Read the article

  • Localized database for customers

    - by Jim
    The company I work for, has just moved to AWS - and currently they have one very large central database with the instance currently located in America. However, one of their clients has requested that all of their data is held in the EU. So creating an AWS instance in Ireland isn't a problem, the problem is the database and how to manage it. We were considering having another database that runs in the EU for European customers, and use a different primary key step, so that the primary keys will never conflict in case the two locations need to be merged in the future. The problem is, if we have a customer that uses our system in both America, and the EU we would have to create 2 accounts for that user, and reporting across both regions would not be possible as the connection time would be too high. Is there an alternative to set this up?

    Read the article

  • How do ulimit -n and /proc/sys/fs/file-max differ?

    - by bantic
    I notice that on a new CentOS image that I just booted up off of EC2 that the ulimit default is 1024 open files, but /proc/sys/fs/file-max is set at 761,408 and I'm wondering how these two limits work together. I'm guessing that ulimit -n is a per-user limit of number of file descriptors while /proc/sys/fs/file-max is system-wide? If that's the case, say I've logged in twice as the same user -- does each logged-in user have a 1024 limit on number of open files, or is it a limit of 1024 combined open files between each of those logged-in users? And is there much performance impact to setting your max file descriptors to a very high number, if your system isn't ever opening very many files?

    Read the article

  • new pc..noisy fan

    - by BRQ
    It's a new build, but it's always had noisy fans. From start to end, they will not stop running. The case is a cooler master which I believe comes with a fan that is not controlled by BIOS (according to technician), so that may be the source of the problem..but my lack of knowledge on the matter prevents me from making a reasonable assessment. Here are readings from CoreTemp: Model: Intel Core i7 870 (Lynnfield) Platform: LAG 1156 (Socket H) Frequency: 1658.23MHz (132.66 x 12.5) Tj. Max: 99 C Core #0: low= 34 C; high= 42 C; Load= 0% Core #1: low 31 C; high 42 C; load= 0% Core #3: 35 C; 42 C; 0% Any input will be appreciated.

    Read the article

  • Macbook Pro (SantaRosa) internal display not detected by graphics card, external monitor OK

    - by BLAU
    My MacBook Pro (2.4 Ghz, Santa Rosa with infamous nVidia card) acts strange. It shows the normal gray screen with Apple logo and animation flawlessly during start up but the internal display goes black without any rendering at all when all is loaded. (shining a light on display show nothing) If an external monitor is connected through the DVI port it will remain black during start up and then show the desktop as the internal display goes black. This happens both while booting to Mountain Lion and Windows XP. I have checked "About my Mac" and only the external display is listed. The same is the case if I use the nVidia Control panel in Windows XP. My questions: Is this a hardware problem or is it related to software maybe even firmware? What controls the display during start up, graphics card or something else?

    Read the article

  • Backing up Windows 2003 Server that has Certificate Authority

    - by Dina
    I want to export and migrate a Certificate Authority CA role from a Windows 2003 machine to a new copy of Windows 2008 R2 virtual machine. I was told that I cannot have 2 CA roles on the same network at the same time. Therefore, I must first export the certificates on the older machine, delete the CA role, then add the CA role on the new machine and import the certificates into it. As a safety precaution, I am tasked to find a backup solution in case this does not work and I need to revert back to the old Windows 2003 CA. My question is: What is the best software for doing this type of backup? I am currently trying out Symantec Backup Exec 2012. Which I hope will allow me to create a backup prior to removing CA role on Windows 2003. If this CA migration fails, the backup will allow me to revert the old machine to a time before I removed its CA role.

    Read the article

  • Difference between tc qdisc and netdev_max_backlog

    - by Mediocre Gopher
    I'm wondering what the difference between these two things on linux is. According to the docs tc qdisc can be used to set the queue size for egress and ingress packets coming in and out of the NIC (or that's how I understood it). But from what I understand netdev_max_backlog can also be used to set this. If I were to set both of them which would be used? Or are there actually two queues that are being manipulated in this case? If there are two queues, which queue is above the other (if the application is at the top and the hardware at the bottom)?

    Read the article

  • Colored blocks on boot

    - by stackzerad
    When my laptop tries to boot right after POST I see colored blocks with flashing symbols in them. I am able to boot from windows PE cd. Tried fixboot and fixmbr with no success. I have also tried replacing boot files (ntldr, io.sys etc..) and removing video card drivers from windows\system32\drivers. The drive is seagate 2.5 ATA 160GB and has one NTFS partition on it. I have already fixed this issue by reformating the drive and reinstalling everything but after couple of weeks I get the same issue again. The diagnostics software shows no bad sectors on it and virus scan didn't find anything. Does anybody have an idea what this might be? UPDATE: tried defragmenting the hard drive just in case, but still no luck

    Read the article

  • How to revert to "last known good configuration"

    - by Ripley
    Hi Guys. I failed to install ubuntu 10.04 with WUBI, for some reason it's showing me the root partion is not defined. I'm bored to fight with it so I just removed ubuntu in windows. However this installation made my original Windows XP cripple, a normal boot will end up with a blue screen, error code 7E, I'm still able to boot with the 'last known good configuration' tho. My understanding is booting like this will recover things and I'm supposed to be good when reboot, while this is not the case for me, I have to choose the 'boot from last known good configuration' each and every time to work around the blue screen. Could you suggest how could I resolve this? I feel it's foolish having to waste 10 more seconds each time starting the OS.

    Read the article

  • Logon script does not run for all users

    - by Herohtar
    We have a standalone common-use workstation running Windows XP Pro SP3 and have created a script using Javascript (scriptname.js) that is to be run for each user. The file was added as a user logon script via gpedit.msc and tested using a newly created user account as well as an existing user account. The script ran and functioned as intended on both accounts; however, a few existing users have informed us that the script is not running on their accounts. All user accounts are members of the same groups and have identical permissions. We already have an existing script (but in this case, a batch file) that is applied in the same manner and it runs for all accounts without any problems. Furthermore, on the accounts where the new script does not run during logon, it can still be run manually and works fine. So the question is: what would cause this script to not run during logon on certain accounts? Thanks!

    Read the article

  • How can I get cross-browser consistent behavior for TR heights within a table with a set height? [migrated]

    - by Dan
    I have an arbitrary number of tables with an arbitrary number of rows in each, and all tables are the same height. My initial approach was to just set the overall height of the table and hope the rows were smart enough to distribute themselves appropriately. That's not the case. I have 4 different behaviors going on with 4 browsers, but I need them to all render at the very least in a similar way. Safari & Chrome (WebKit): All rows are equal height, creating scroll bars as needed and fitting within table height. Firefox: All rows are the height necessary to fit their content, with the remaining rows overflowing out of the table. Additionally, If the content of the rows does not take up all of the height, only the part of the table with content in it takes the background (though it seems, through use of Firebug, that the actual table [and TR] extend to the bottom of the proper table height). IE: All rows are the height necessary to fit their content, with the remaining rows overflowing out of the table. Obviously this only includes one version of each browser and additional variation would likely appear with more being tested. Ideally, a solution where the browser renders TRs with less content smaller than those with larger content, while still using scrolling within the variable height TRs when the overall height of the table is not enough would be optimum. I could potentially see a solution to achieve that with JS, but can it be done with CSS? Or, if not, can the behavior that WebKit displays be made to work across the browsers? Thanks! PS: Example can be found here.

    Read the article

  • How to setup a tunneled connection to a NAS?

    - by Jan
    Hi there, While looking into alternatives for Dropbox I explored the idea of a NAS. It would be easy to hook up the NAS to a broadband connection and have the users access the files. However, - The data needs to be encrypted on the NAS itself (in case it gets stolen). - The connection to the NAS needs to be secure (VPN?). - The drive needs to appear as a drive under windows (so no https). Any ideas? Thanks in advance, I can't seem to find the final solution! Jan

    Read the article

  • CoffeeScript - inability to support progressive adoption

    - by Renso
    First if, what is CoffeeScript?Web definitionsCoffeeScript is a programming language that compiles statement-by-statement to JavaScript. The language adds syntactic sugar inspired by Ruby and Python to enhance JavaScript's brevity and readability, as well as adding more sophisticated features like array comprehension and pattern matching.The issue with CoffeeScript is that it eliminates any progressive adoption. It is a purist approach, kind of like the Amish, if you're not borne Amish, tough luck. So for folks with thousands of lines of JavaScript code will have a tough time to convert it to CoffeeScript. You can use the js2coffee API to convert the JavaScript file to CoffeeScript but in my experience that had trouble converting the files. It would convert the file to CoffeeScript without any complaints, but then when trying to generate the CoffeeScript file got errors with guess what: INDENTATION!Tried to convince the CoffeeScript community on github but got lots of push-back to progressive adoption with comments like "stupid", "crap", "child's comportment", "it's like Ruby, Python", "legacy code" etc. As a matter of interest one of the first comments were that the code needs to be re-designed before converted to CoffeeScript. Well I rest my case then :-)So far the community on github has been very reluctant to even consider introducing some way to define code-blocks, obviously curly braces is not an option as they use it for json object definitions. They also have no consideration for a progressive adoption where some, if not all, JavaScript syntax will be allowed which means all of us in the real world that have thousands of lines of JavaScript will have a real issue converting it over. Worst, I for one lack the confidence that tools like js2coffee will provide the correct indentation that will determine the flow of control in your code!!! Actually it is hard for me to find enough justification for using spaces or tabs to control the flow of code. It is no wonder that C#, C, C++, Java, all enterprise-scale frameworks still use curly braces. Have never seen an enterprise app built with Ruby or PhP.Let me know what your concerns are with CoffeeScript and how you dealt with large scale JavaScript conversions to CoffeeScript.

    Read the article

  • Don't miss Virtual Developer Day - All about ADF next week

    - by Shay Shmeltzer
    In case you haven't heard we are holding a free online virtual developer day next week - July 10th that you should attend - even if you think you already know ADF. First the registration link - http://bit.ly/fusiondev. While one of the tracks is aimed at developer who are relatively new to ADF - and cover ADF Faces, ADF Controller and a comparison of productivity with Forms and other tools - the two other tracks have great content on some topics that you might not be familiar with even if you already work with ADF. This include sessions about the upcoming ADF Mobile, The new ADF support in Eclipse and information about Application Life Cycle Management with ADF and JDeveloper. As well as sessions that will open your mind to the areas where ADF integrates with other Fusion Middleware Solutions such as ADF integration with BI, WebCenter and SOA. Most of the sessions are quite heavy on demos and you'll get a chance to interact with the presenters and ask questions during the live event. You should register even if you can't attend the live event - this way you'll get an email pointing you to the recorded sessions for on demand viewing. See you next week.

    Read the article

< Previous Page | 416 417 418 419 420 421 422 423 424 425 426 427  | Next Page >