Search Results

Search found 23021 results on 921 pages for 'process monitoring'.

Page 47/921 | < Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >

  • Multi MVC processing vs Single MVC process

    - by lordg
    I've worked fairly extensively with the MVC framework cakephp, however I'm finding that I would rather have my pages driven by the multiple MVC than by just one MVC. My reason is primarily to maintain an a more DRY principle. In CakePHP MVC: you call a URL which calls a single MVC, which then calls the layout. What I want is: you call a URL, it processes a layout, which then calls multiple MVC's per component/block of html on the page. When you compare JavaScript components, AJAX, and server side HTML rendering, it seems the most consistent method for building pages is through blocks of components or HTML views. That way, the view block could be situated either on the server or the client. This is technically my ONLY disagreement with the MVC model. Outside of this, IMHO MVC rocks! My question is: What other RAD frameworks follow the same principles as MVC but are driven rather by the View side of MVC? I've looked at Django and Ruby on Rails, yet they seems to be more Controller driven. Lift/Scala appears to be somewhat of a good fit, but i'm interested to see what others exist.

    Read the article

  • Monitoring outgoing bandwidth of application

    - by jnolte
    I currently have a VPS that is consuming a ton of outgoing bandwidth and I am trying to drill down to where this may be coming from. Does anyone know of a logical way to go about finding out which pages on the site are consuming the most outgoing data. We have done a ton of front-end optimizations to the site and our google page speed rankings ar 85% so I feel we have done a pretty great job at optimizing the site for speed. Can someone lend some insight on how they have made similar optimizations? Application / Server Stack LEMP Running Varnish Cache / PHP5-FPM WordPress running w3 Total Cache Ubuntu 12.04 LTS

    Read the article

  • Using awstats with a round-robin DNS configuration

    - by Shaun
    I have a website with multiple web servers whose access is controlled via a round-robin DNS. We currently use Google Analytics for site traffic monitoring but were looking to move to awstats due to concerns of inaccuracy with Google Analytics and using third-party trackers in general. I have a little experience with awstats and I know it gets its information from parsing server logs. How would this work when you have multiple web servers logging independently to separate locations? Is this supported with awstats? Is there an alternative I could use to track traffic activity directly on my servers?

    Read the article

  • Process, Participate, Play: Oracle BPM and SOA at Oracle OpenWorld

    - by Oracle OpenWorld Blog Team
    Oracle OpenWorld 2012 provides a unique opportunity for BPM and SOA professionals to meet industry leaders and peers, and get insight into the latest product advancements that will help their companies gain a competitive advantage.Via a variety of sessions, hands-on labs, birds-of-a-feather sessions, and demos, attendees will learn how Oracle SOA Suite, Oracle BPM Suite, and Oracle SOA Governance provide a unified and collaborative environment for design and deployment of dynamic business processes. Topics include architecture, integration, implementation, and best practices for on-premises or cloud deployments. Participants will learn how new capabilities of BPM and SOA can help their enterprises gain unprecedented visibility, agility and efficiencies.Maximize the value of attending Oracle Open World by attending sessions that best meet your needs and goals. This exciting series of SOA and BPM sessions is focused on three different audience segments. Business managers or business analysts, click here  IT executives or enterprise architects, click here Developers looking to sharpen their SOA skills, click here To stay in touch with the details and announcements for Oracle BPM Suite and Oracle SOA Suite, check out the BPM and SOA blogs.

    Read the article

  • Find out which task is generating a lot of context switches on linux

    - by Gaks
    According to vmstat, my Linux server (2xCore2 Duo 2.5 GHz) is constantly doing around 20k context switches per second. # vmstat 3 procs -----------memory---------- ---swap-- -----io---- -system-- ----cpu---- r b swpd free buff cache si so bi bo in cs us sy id wa 2 0 7292 249472 82340 2291972 0 0 0 0 0 0 7 13 79 0 0 0 7292 251808 82344 2291968 0 0 0 184 24 20090 1 1 99 0 0 0 7292 251876 82344 2291968 0 0 0 83 17 20157 1 0 99 0 0 0 7292 251876 82344 2291968 0 0 0 73 12 20116 1 0 99 0 ... but uptime shows small load: load average: 0.01, 0.02, 0.01 and top doesn't show any process with high %CPU usage. How do I find out what exactly is generating those context switches? Which process/thread? I tried to analyze pidstat output: # pidstat -w 10 1 12:39:13 PID cswch/s nvcswch/s Command 12:39:23 1 0.20 0.00 init 12:39:23 4 0.20 0.00 ksoftirqd/0 12:39:23 7 1.60 0.00 events/0 12:39:23 8 1.50 0.00 events/1 12:39:23 89 0.50 0.00 kblockd/0 12:39:23 90 0.30 0.00 kblockd/1 12:39:23 995 0.40 0.00 kirqd 12:39:23 997 0.60 0.00 kjournald 12:39:23 1146 0.20 0.00 svscan 12:39:23 2162 5.00 0.00 kjournald 12:39:23 2526 0.20 2.00 postgres 12:39:23 2530 1.00 0.30 postgres 12:39:23 2534 5.00 3.20 postgres 12:39:23 2536 1.40 1.70 postgres 12:39:23 12061 10.59 0.90 postgres 12:39:23 14442 1.50 2.20 postgres 12:39:23 15416 0.20 0.00 monitor 12:39:23 17289 0.10 0.00 syslogd 12:39:23 21776 0.40 0.30 postgres 12:39:23 23638 0.10 0.00 screen 12:39:23 25153 1.00 0.00 sshd 12:39:23 25185 86.61 0.00 daemon1 12:39:23 25190 12.19 35.86 postgres 12:39:23 25295 2.00 0.00 screen 12:39:23 25743 9.99 0.00 daemon2 12:39:23 25747 1.10 3.00 postgres 12:39:23 26968 5.09 0.80 postgres 12:39:23 26969 5.00 0.00 postgres 12:39:23 26970 1.10 0.20 postgres 12:39:23 26971 17.98 1.80 postgres 12:39:23 27607 0.90 0.40 postgres 12:39:23 29338 4.30 0.00 screen 12:39:23 31247 4.10 23.58 postgres 12:39:23 31249 82.92 34.77 postgres 12:39:23 31484 0.20 0.00 pdflush 12:39:23 32097 0.10 0.00 pidstat Looks like some postgresql tasks are doing 10 context swiches per second, but it doesn't all sum up to 20k anyway. Any idea how to dig a little deeper for an answer?

    Read the article

  • Whatsup gold reporting

    - by rotard
    I work for an IT consulting company, and we use WhatsUp Gold to monitor our and our clients' servers. WUG works, and provides all the data we need in a reasonably useful format. However, management wants to use it as a sales tool. They want more attractive reports that they can hand off to clients and prospective clients to show off how slick our monitoring software is. Is there a third-party solution that will generate more attractive reports based on data in the WUG database? I have already suggested using reporting services to build our own reports, but that idea was rejected.

    Read the article

  • BixData or Zabbix?

    - by Arafat
    Hi all, I've been using Ganglia to monitor my single Mac OSX server which runs Apache and MySQL. I'm ok with it. Now we are upgrading our servers, 6 IBM X3650 M3 and 2 Fujitsu servers. 2 IBM for Apache cluster and 4 IBM for MySQL NDB Cluster. The other two servers are for Load balancers. All servers are going to run Debian Lenny 5 on it. Now I need to decide on which monitoring tool I should go for. I found that BixData and Zabbix does an excellent job than Ganglia, in terms of sensors and reporting. Have anyone tried the above two tools? And which tool would you suggest me? For Debian. As I'm writing this, I'm installing BixData to try.... Thanks in advance.

    Read the article

  • Avoid GPL violation by moving library out of process

    - by Andrey
    Assume there is a library that is licensed under GPL. I want to use it is closed source project. I do following: Create small wrapper application around that GPL library that listens to socket, parse messages and call GPL library. Then returns results back. Release it's sources (to comply with GPL) Create client for this wrapper in my main application and don't release sources. I know that this adds huge overhead compared to static/dynamic linking, but I am interested in theoretical way.

    Read the article

  • Problem with monitoring Glassfish with JConsole..

    - by Pavel
    I have enabled JMX connection on remote Glassfish server and then I've restarted it. During starting server notified: Standard JMX Clients (like JConsole) can connect to JMXServiceURL: [service:jmx:rmi:///jndi/rmi://myserver:8686/jmxrmi] for domain management purposes. Port 8686 is opened for connections. But I can't connect to server with JConsole.. It says: Connection failed. How can I solve this problem? Thanks in advance.

    Read the article

  • Wireless Activity Monitoring for PCI DSS Compliance

    - by dkusleika
    In an effort to be PCI DSS compliant, I took a trustkeeper.net questionnaire. I failed the question that asks Is the presence of wireless access points tested for by using a wireless analyzer at least quarterly or by deploying a wireless IDS/IPS to identify all wireless devices in use? (SAQ #11.1) My only wireless access point is outside my firewall, so even if you cracked my wireless you couldn't get inside my domain (unless you crack that too). My firewall doesn't have IPS and I couldn't tell if it had IDS. I looked around for a wireless analyzer, but what I found was $500, which is a little pricey for my size business. And even if I got it, I'm not sure I would understand what it tells me. Surely there are smaller/less sophisticated businesses that take credit cards and have solved this. My questions are: What are the risks if someone were to crack my wireless? (Could they read all internet traffic? Just wireless traffic? Just use my internet connection?) And what is the best/cheapest way to test my connection point quarterly? Should I buy the $500 analyzer? Domain is Windows Server 2000. Firewall is Sonicwall Pro 2040. Router is 8 port D-link.

    Read the article

  • Printer monitoring script (PowerShell)

    - by HannesFostie
    I am going to write a script of some sort to check event viewer in a windows server 2003 for all printjobs, and then write them to a comma delimited textfile like printername_floor_room.txt I am wondering what the best way is to do this realtime, and keep checking the event viewer constantly. Any caveats I need to be aware of? Thanks EDIT: Okay, so I will most likely go for PowerShell and use Get-EventLog and then edit the "table" data. Problems I'm having: if I were to save all this data to a text file, how do I get the data out of it? A comma-separated file I could work with, but this, I'm not really sure. And once that is sorted out, I'm still not sure how to keep the file updated more or less real-time. Can I make this service-like, without hogging up all resources? Run it every x seconds for example?

    Read the article

  • How can I find the program making a harmonica sound?

    - by Josh
    A friend has a Windows XP SP3 machine that plays a harmonica sound for about 5 seconds throughout the day at what seems to be random intervals (every couple hours). My question is how can I find the program making this sound? Is there a Windows API hook for monitoring audio access? I've gone through and checked all the standard Windows sounds in the Control Panel and right now the theme is set to no sounds and I personally checked to make sure none of the events have a sound specified. I also checked the Task Scheduler to make sure there wasn't something scheduled to go off every couple hours. Any ideas on how to go about finding the bugger?

    Read the article

  • Unable to Get IIS Response Times with Hyperic Monitoring Tool

    - by jwmajors81
    We have a very large .NET application (vendor) that I am trying to gather performance metrics for using Hyperic. In general I wanted to be able to report response times for each of the components within the application, which include: Web Services ASP.NET Pages MSMQ I am currently unable to successfully monitor the response times for IIS on my windows machines. I have successfully auto-discovered them and I am getting diagnostic information, but am not getting back the response times. After looking online at what other people are seeing, I found that I am missing a tab for the response times which is usually next to the metrics tab. Also, when I look at the configuration screen for IIS I do not see a field which enables me to specify where the log files are located at. Please note that my logs are located at e:\LogFiles\ instead of the usual location because our IT staff doesn't allocate much space for the C: drive. Please note that I have the MSMQ monitors up and running great. Any assistance would be greatly appreciated. Thanks, Jeremy

    Read the article

  • Monitoring IIS with SNMP

    - by Shtong
    Hello, I'm trying to monitor IIS using the Windows integrated SNMP service (for versions 2000, 2003 ans 2008 of Windows Server, but i'm doing my experiments on a 2008 R2 box). So far I was able to install the SNMP service, and query mib-2 values without problems. Now, I'm having trouble to monitor IIS informations, as the SNMP service does not seem to know about the .iso.org.dod.internet.private.enterprises.microsoft.software.internetServer OIDs. Walking the entire SNMP tree only gets the mib-2 stuff as well as the names of some services, but that's it. I did't find anything interesting in the service configuration panel. So how do I configure this service to get IIS information, or other WMI data for that matter (I installed the SNMP WMI Provider) ? I can't find any documentation about that !

    Read the article

  • Monitoring outgoing messages using EXIM

    - by dashmug
    I work as an IT guy in a law firm. I am recently asked to make a system wherein all the outgoing emails coming from our server to our clients will be put on hold first and wait for approval before it gets sent to the client. Our mail server uses Exim (that's what it says in cPanel). I am planning to create filters where the outgoing emails will be forwarded to an editor account. Then, the editor will review and edit the contents of the email. When the editor already approves the email, it will then get sent to the client by the editor but still using the original sender in the "From:" and "Reply-To:" field. I found some pointers from this site = http://www.devco.net/archives/2006/03/24/saving_copies_of_all_email_using_exim.php. Once the filters are in place, I want to make a simple PHP interface for the editor to check the forwarded emails and edit them if necessary. The editor can then click on an "Approve" button that will finally deliver the message using the original sender. I'm also thinking that maybe a PHP-less system will be enough. The editor can receive the emails from his own email client edit them and simply send the email as if he is the original sender. Is my plan feasible? Will there be issues that I have overlooked? Does it have the danger of being treated as spam by the other mailservers since I'll be messing up the headers?

    Read the article

  • Apache Process question about RAM usage

    - by Andrew Fashion
    So everytime I load a new page, I notice a new HTTPD process opens, every time I click a page, and each process says it's using anywhere from 2-4.5% of memory. Does that mean every single process is running at that time using 2-4% of RAM? It's a brand new server and I'm the only one on the server at the moment. Or does it mean all the other processes are dying, and only the new one is active. Because 4% of my 2048MB of RAM is already 82MB for just one process!?!? Let me know, because I am trying to determine what I need to beef my server up in order to handle high loads of traffic. I'm expect to get 20,000 uniques per day on launch. I am currently running a Dual Quad Xeon server, with only 2GB of ram, I will upgrade to 8GB or more shortly. Let me know what you suggest! thank you [root@D18634 log]# top | grep 'httpd' 11315 apache 15 0 362m 82m 24m R 12.3 4.1 0:03.00 httpd 11310 apache 16 0 322m 41m 21m S 5.7 2.1 0:02.98 httpd 11315 apache 15 0 362m 83m 25m S 24.3 4.1 0:03.73 httpd 11319 apache 16 0 324m 42m 20m R 1.0 2.1 0:01.85 httpd 11319 apache 16 0 362m 82m 23m R 78.5 4.1 0:04.21 httpd 11321 apache 16 0 323m 44m 23m S 35.3 2.2 0:04.13 httpd 11319 apache 15 0 361m 82m 23m S 8.3 4.1 0:04.46 httpd 11321 apache 15 0 323m 44m 23m S 35.9 2.2 0:05.21 httpd 11313 apache 15 0 324m 41m 19m S 48.6 2.1 0:03.23 httpd 11322 apache 16 0 354m 72m 20m R 11.0 3.6 0:05.11 httpd 11322 apache 16 0 354m 72m 20m S 23.9 3.6 0:05.83 httpd 11314 apache 16 0 355m 75m 22m R 18.3 3.7 0:04.64 httpd

    Read the article

  • why is mailgraph not updating ?

    - by jack
    I just installed mailgraph on a Linux Ubuntu 10.04 box monitoring Postifix. Everything seems to be working fine, but the graphs are not always updating. When I refresh the web page or close the browser and come back later, the graph updates and moves to the left, showing new time, the the data line is not drawn. Sometimes it does not update for several hours. I had started, stopped and restarted mailgraph-init and that doesn't seem to make a difference. Any ideas?

    Read the article

  • Tips/Process for web-development using Django in a small team

    - by Mridang Agarwalla
    We're developing a web app uing Django and we're a small team of 3-4 programmers — some doing the UI stuff and some doing the Backend stuff. I'd love some tips and suggestions from the people here. This is out current setup: We're using Git as as our SCM tool and following this branching model. We're following the PEP8 for your style guide. Agile is our software development methodology and we're using Jira for that. We're using the Confluence plugin for Jira for documentation and I'm going to be writing a script that also dumps the PyDocs into Confluence. We're using virtualenv for sandboxing We're using zc.buildout for building This is whatever I can think of off the top of my head. Any other suggestions/tips would be welcome. I feel that we have a pretty good set up but I'm also confident that we could do more. Thanks.

    Read the article

  • Cannot connect to Solaris Server when Oracle GoldenGate process uses the port

    - by Abdallah Ghrb
    I'm trying to test the Oracle goldengate to replicate data between two databases on the same server. So I installed two databases and two goldengate homes in the same machine. The goldengate processes are started from each home and they are responsible for the replication : Process from home 1 configured on port 7809 & process for home 2 configured on port 7810. For a successful replication, processes started from goldengate home 1 should communicate with processes started from goldengate home 2. But for some reasons, this is not happening. The goldengate log file has the following error : OGG-01223 Oracle GoldenGate Capture for Oracle, exthrr.prm: TCP/IP error 131 (Connection reset by peer). "Googling" for this error, it said that the connection occurred but the host terminated it. Tried to telnet the machine with the used port and it gave the following error: bash-3.00$ telnet 10.10.3.124 7810 Trying 10.10.3.124... Connected to 10.10.3.124. Escape character is '^]'. Connection to 10.10.3.124 closed by foreign host. Here the communication occurs but for only around 3 seconds then it is closed by the host, which is the same explanation of the above error in the goldengate log file. I tried to change the port but still the same error. The problem is happening only when the goldengate process is using the port. When other process is using the same port, I can telnet successfully.

    Read the article

  • zabbix monitoring mysql database

    - by krisdigitx
    I have a server running multiple instances of mysql and also has the zabbix-agent running. In zabbix_agentd.conf i have specified: UserParameter=multi.mysql[*],mysqladmin --socket=$1 -uzabbixagent extended-status 2>/dev/null | awk '/ $3 /{print $$4}' where $1 is the socket instance. From the zabbix server i can run the test successfully. zabbix_get -s ip_of_server -k multi.mysql[/var/lib/mysql/mysql2.sock] and it returns all the values However the zabbix item/trigger does not generate the graphs, I have created a MACRO for $1 which is the socket location {$MYSQL_SOCKET1} = '/var/lib/mysql/mysql2.sock' and i use this key in items to poll the value multi.mysql[{$MYSQL_SOCKET1},Bytes_sent] LOGS: this is what i get on the logs: 3360:20120214:144716.278 item [multi.mysql['/var/lib/mysql/mysql2.sock',Bytes_received]] error: Special characters '\'"`*?[]{}~$!&;()<>|#@' are not allowed in the parameters 3360:20120214:144716.372 item [multi.mysql['/var/lib/mysql/mysql2.sock',Bytes_sent]] error: Special characters '\'"`*?[]{}~$!&;()<>|#@' are not allowed in the parameters Any ideas where the problem could be? FIXED {$MYSQL_SOCKET1} = /var/lib/mysql/mysql2.sock i removed the single quotes from the line and it worked...

    Read the article

  • monitoring changes in windows of screen while screen is detached

    - by gojira
    It is possible to monitor windows in screen (I mean the terminal multiplexer called screen) with Ctrl-a M. However, I am only aware of the notification when I am looking at any window. What I want, though, is to somehow also be notified if the screen in question has been detached with Ctrl-a d. I.e., I issue the command to monitor a window in screen, then detach that screen, now I want to get a notification if the monitor detects activity, in some form (a string in the bash I'm using, or an email, or anything). Is this possible, and if yes how?

    Read the article

  • Homebrewed Headphone Amp Shows Off DIY Resin Casting Process

    - by Jason Fitzpatrick
    Although the guts of this DIY build are worth checking out in their own right, what really caught our eye was the beautiful resin cast case surrounding the build as it’s something that could be adapted to a wide variety of projects. Over at the electronics blog Runaway Brainz they were looking for a slick way to encase an amp project. Rather than go with a project box or similar construction, they did a resin pour and then sanded and polished the resulting encasement. The results are stunning and turn the electronic guts of the amp into a work of art. Freeform Headphone Amp [via Make] How to Stress Test the Hard Drives in Your PC or Server How To Customize Your Android Lock Screen with WidgetLocker The Best Free Portable Apps for Your Flash Drive Toolkit

    Read the article

  • Supervisor VS cronjob

    - by Guandalino
    Actually I'm using supervisor to monitor a process and restart it when it stops for some reason. The problem is that in case of a supervisor crash the process stops get monitored. So I thought to schedule a cronjob to check supervisor is running, and eventually restart it. The next thing I'm considering is to get rid of supervisor and check my process directly from the cronjob. I read that sometimes supervisor uses too much memory (to be verified, though). What are the pros in having supervisor VS cronjob monitoring the process?

    Read the article

  • Process Oracle OER Events using a simple Web Service

    - by Bob Webster
    This post provides an example of a simple web service that processes Oracle Enterprise Repository (OER) Events. The service receives events from OER and utilizes the OER REX API to implement simple OER automations for selected event types. The web service example implements the following: When a new Asset is Submitted to OER: The Asset is automatically Accepted by a defined user. When an Asset is Accepted: The Asset is automatically assigned  to a defined user for review. If the accepted asset is of type Service The Version meta data attribute is set based on the version id contained in the suffix of the Service Namespace.      When an Asset is Registered: If the registered Asset is of type Service The related Assets ( Interface and Endpoint are also automatically registered. The sample web service is not intended to replace the out of the Box OER BPM Based workflows, but the service can be utilized in cases where only simple automation is required and the developer has a Java skill set. The service is a lightweight web application that can be easily deployed to the same server as OER or on a different server. Read the complete post here

    Read the article

  • monitoring TCP/IP performance on Solaris

    - by Andy Faibishenko
    I am trying to tune a high message traffic system running on Solaris. The architecture is a large number (600) of clients which connect via TCP to a big Solaris server and then send/receive relatively small messages (.5 to 1K payload) at high rates. The goal is to minimize the latency of each message processed. I suspect that the TCP stack of the server is getting overwhelmed by all the traffic. What are some commands/metrics that I can use to confirm this, and in case this is true, what is the best way to alleviate this bottleneck?

    Read the article

< Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >