Search Results

Search found 32035 results on 1282 pages for 'model view'.

Page 496/1282 | < Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >

  • Imperative vs. component based programming [closed]

    - by AlexW
    I've been thinking about how programming and more specifically the teaching of programming is advocated amongst the community (online). Often I've heard that Ruby and RoR is an ideal platform for learning to program. I completely disagree... RoR and Ruby are based on the application of the component based paradigm, which means they are ideal for rapid application development. This is much like the MVC model in PHP and ASP.NET But, learning a proper imperative language like Java or C/C++ (or even Perl and PHP) is the only way for a new programmer to explore logic itself, and not get too bogged down in architectural concerns like the need for separation of concerns, and the preference for components. Maybe it's a personal preference thing. I rather think that the most interesting aspects to programming are the procedural bits of code I write that actually do stuff rather than the project planning, and modelling that comes about from fully object oriented engineering or simply using the MVC model. I know this may sound confused to some of you. I feel strongly though that the best way for programming to be taught is through imperative and procedural methods. Architectural (component) methods come later, if at all. After all, none of the amazing algorithms that exist were based on OOP practice! It's all procedural code when it comes to the 'magic'. OOP is useful in creating products and utilities. Algorithms are what makes things happen, and move data around, and so imperative (and/or procedural) code are what matters most. When I see programmers recommending Ruby on Rails to newbie developers, I think it's just so wrong. Just because you write less code with Ruby does not make it easier to do! It's the opposite... you have to know loads more to appreciate its succinct nature. New coders who really want to understand the nuts and bolts of coding need to go away and figure out writing methods/functions (i.e. imperative programming) and working in procedural style, in order to grasp the fundamentals, first, before looking into architectural ways of working. So, my question is: should Ruby ever be recommended as a first language? I think no (obviously)... what arguments are there for it?

    Read the article

  • New HP dm4 - No audio on ubuntu 11.10 64bits

    - by Haze1
    I just got a new laptop, HP dm4, and I'm having problems getting the audio to work properly on it. http://www.alsa-project.org/db/?f=7b697a35465a9f7236fb94deb9ff97fa65e55489 I tried to edit /etc/modprobe.d/alsa-base.conf and added: option snd-hda-intel model=ref this caused the audio to work, but it's muffled. I'm wondering if anybody knows what would be the correct options to get this POS to work. Thanks in advance

    Read the article

  • Why is my concurrency capacity so low for my web app on a LAMP EC2 instance?

    - by AMF
    I come from a web developer background and have been humming along building my PHP app, using the CakePHP framework. The problem arose when I began the ab (Apache Bench) testing on the Amazon EC2 instance in which the app resides. I'm getting pretty horrendous average page load times, even though I'm running a c1.medium instance (2 cores, 2GB RAM), and I think I'm doing everything right. I would run: ab -n 200 -c 20 http://localhost/heavy-but-view-cached-page.php Here are the results: Concurrency Level: 20 Time taken for tests: 48.197 seconds Complete requests: 200 Failed requests: 0 Write errors: 0 Total transferred: 392111200 bytes HTML transferred: 392047600 bytes Requests per second: 4.15 [#/sec] (mean) Time per request: 4819.723 [ms] (mean) Time per request: 240.986 [ms] (mean, across all concurrent requests) Transfer rate: 7944.88 [Kbytes/sec] received While the ab test is running, I run VMStat, which shows that Swap stays at 0, CPU is constantly at 80-100% (although I'm not sure I can trust this on a VM), RAM utilization ramps up to about 1.6G (leaving 400M free). Load goes up to about 8 and site slows to a crawl. Here's what I think I'm doing right on the code side: In Chrome browser uncached pages typically load in 800-1000ms, and cached pages load in 300-500ms. Not stunning, but not terrible either. Thanks to view caching, there might be at most one DB query per page-load to write session data. So we can rule out a DB bottleneck. I have APC on. I am using Memcached to serve the view cache and other site caches. xhprof code profiler shows that cached pages take up 10MB-40MB in memory and 100ms - 1000ms in wall time. Pages that would be the worst offenders would look something like this in xhprof: Total Incl. Wall Time (microsec): 330,143 microsecs Total Incl. CPU (microsecs): 320,019 microsecs Total Incl. MemUse (bytes): 36,786,192 bytes Total Incl. PeakMemUse (bytes): 46,667,008 bytes Number of Function Calls: 5,195 My Apache config: KeepAlive On MaxKeepAliveRequests 100 KeepAliveTimeout 3 <IfModule mpm_prefork_module> StartServers 5 MinSpareServers 5 MaxSpareServers 10 MaxClients 120 MaxRequestsPerChild 1000 </IfModule> Is there something wrong with the server? Some gotcha with the EC2? Or is it my code? Some obvious setting I should look into? Too many DNS lookups? What am I missing? I really want to get to 1,000 concurrency capacity, but at this rate, it ain't gonna happen.

    Read the article

  • Munin help... configure graphs to show different time periods

    - by PiZzL3
    I'm running a server with Munin installed through WHM cPanel. I've been googling around and I can't seem to find out how to do the following: A: Change a graph to show a different time period (currently I can only view, day, week, month, year). I would like to possibly view per minute, per hour... or specific intervals such as per 30 minutes, per 4 hours, etc... B(optional, but preferred): Add new graphs with the above criteria(A) above. I am a novice with all things that will possibly be required to do this (such as ssh). I wish I could just go edit a php file as I'm highly experienced in that subject. Thanks for any help!

    Read the article

  • Role based access to resources for a RESTful service

    - by mutex
    I'm still wrapping my head around REST, but I wonder if someone can help with any suggestions or approaches to role based access control for a RESTful service, particularly from the point of view of securing the data and how the URLs might look. It's probably best to consider an example: Say I have a REST service for Customers, and want to split the users of this REST service into Admin, Editor and Reader roles: Admins can change all attributes of a Customer resource Editors can change only some Readers can only view them. Access control rights are assigned to the Customers entities individually. So for example a user of the service might have admin rights to Customers 1,2 and 3 but Editor access to 4,5 and Reader access to 7,8,9. Now consider the user calling the service. What is a good way to seperate the list of Customers for the current User? GET /Customer - this might get a list of all customers that the current user has Admin\Editor\Reader access to. But then on each Customer the consumer would need an indication of what role they have. Or would it be "better" having something like GET /Customer/Admin - return all customers the current user has Admin access to. Just looking for some high level pointers or reading on a decent way to secure\filter the resources based on roles of the current user.

    Read the article

  • Creating a Simple ASP.NET Report with Export to Excel

    In this article you will learn how to create a simple ASP.NET report using Web Forms, C#, and a View Model class rather than drag and drop controls, resulting in very clean and understandable HTML. Then, you'll learn how to add Export to Excel functionality, allowing users to export the data in Excel format and save the file with a default filename of your choosing (as opposed to Report.aspx, for instance).

    Read the article

  • AngularJS dealing with large data sets (Strategy)

    - by Brian
    I am working on developing a personal temperature logging viewer based on my rasppi curl'ing data into my web server's api. Temperatures are taken every 2 seconds and I can have several temperature sensors posting data. Needless to say I will have a lot of data to handle even within the scope of an hour. I have implemented a very simple paging api from the server so the server doesn't timeout and is currently only returning data in 1000 units per call, then paging through the data. I had the idea to intially show say the last 20 minutes of data from a sensor (or all sensors depending on user choices), then allowing the user to select other timeframes from which to show data. The issue comes in when you want to view all sensors or an extended time period (say 24 hours). Is there a best practice of handling this large amount of data? Would it be useful to load those first 20 minutes into the live view and then cache into local storage something like the last 24 hours? I haven't been able to find a decent idea of this in use yet even though there are a lot of ways to take this problem. I am just looking for some suggestions as to what might provide a good balance between good performance and not caching the entire data set on the client side (as beyond a week of data this might not be feasible).

    Read the article

  • Why C++ people loves multithreading when it comes to performances?

    - by user1849534
    I have a question, it's about why programmers seems to love concurrency and multi-threaded programs in general. I'm considering 2 main approach here: an async approach basically based on signals, or just an async approach as called by many papers and languages like the new C# 5.0 for example, and a "companion thread" that maanges the policy of your pipeline a concurrent approach or multi-threading approach I will just say that I'm thinking about the hardware here and the worst case scenario, and I have tested this 2 paradigms myself, the async paradigm is a winner at the point that I don't get why people 90% of the time talk about concurrency when they wont to speed up things or make a good use of their resources. I have tested multi-threaded programs and async program on an old machine with an Intel quad-core that doesn't offer a memory controller inside the CPU, the memory is managed entirely by the motherboard, well in this case performances are horrible with a multi-threaded application, even a relatively low number of threads like 3-4-5 can be a problem, the application is unresponsive and is just slow and unpleasant. A good async approach is, on the other hand, probably not faster but it's not worst either, my application just waits for the result and doesn't hangs, it's responsive and there is a much better scaling going on. I have also discovered that a context change in the threading world it's not that cheap in real world scenario, it's infact quite expensive especially when you have more than 2 threads that need to cycle and swap among each other to be computed. On modern CPUs the situation it's not really that different, the memory controller it's integrated but my point is that an x86 CPUs is basically a serial machine and the memory controller works the same way as with the old machine with an external memory controller on the motherboard. The context switch is still a relevant cost in my application and the fact that the memory controller it's integrated or that the newer CPU have more than 2 core it's not bargain for me. For what i have experienced the concurrent approach is good in theory but not that good in practice, with the memory model imposed by the hardware, it's hard to make a good use of this paradigm, also it introduces a lot of issues ranging from the use of my data structures to the join of multiple threads. Also both paradigms do not offer any security abut when the task or the job will be done in a certain point in time, making them really similar from a functional point of view. According to the X86 memory model, why the majority of people suggest to use concurrency with C++ and not just an async aproach ? Also why not considering the worst case scenario of a computer where the context switch is probably more expensive than the computation itself ?

    Read the article

  • Any addin or tool to open multiple web sites on one page?

    - by David.Chu.ca
    Is there any addin or tools available for web browser to view multiple web sites on one page or tab? Or is there any web service to allow me to customize one page for multiple web sites? For example, within a page, there are some tools to allow me to define regions like a table or several frames. Then I'll be able to add url to each cell or frames. After settings, I'll be able to view several sites on one tab or page. Not sure if there is any tools available?

    Read the article

  • OpenGL, objects disappear at a certain z distance

    - by smoth190
    I'm writing a managed OpenGL library in C++, and I'm having a pretty annoying problem. Whenever I set an objects position to -2.0 or lower, the object disappears. And at distances 0 through -1.9, it doesn't appear to move away from them camera. I have a world matrix (which is multiplied by the objects position to move it), a view matrix (which is just the identity matrix currently) and a projection matrix, which is setup like this: FOV: 45.0f Aspect Ratio: 1 zNear: 0.1f zFar: 100.0f using GLMs glm::perspective method. The order the matrices are multiplied by in the shader is world, view, projection, then position. I can move the object along the X and Y axis perfectly fine. I have depth testing enabled, using GL_LEQUAL. I can change the actually vertices positions to anything I want, and they move away from the camera or towards it perfectly fine. It just seems to be the world matrix acting up. I'm using glm::mat4 for the world matrix, and glm::vec3 for positions. Whats going on here? I'm also using OpenGL 3.1, GLSL version 140 (1.4?).

    Read the article

  • Can my machine run Ubuntu( kubuntu | xubuntu | mint ) 12.04 WELL?

    - by Steve
    I have a 9 year old computer packing the hardware listed below. My question is, can I run 12.04 ( Ubuntu, Kubuntu, MINT or Xubuntu ) WELL? I was running Ubuntu 10.10 and I upgraded to 12.04 by going through each release via the update manager: 11.04 - 11.10 -12.04 During the installation process for 12.04 I saw an error message that there was an error installing and setting up part of the kernel. Later, when I tried installing a package in synaptic, I got another error message mentioning the kernel. When I rebooted, I got told somehting about my video and graphics was not configured properly and that I would have to do it manually ( like I know ). It gave me the option to enter the system in low graphics mode, but it just hanged. I had an old livecd of Xubuntu 10.10 around so I used that to get into my computer and copy data over to an external hard drive. I think tried to install Xubuntu 10.10 from the livecd, with the option "download updates" checked. The install process moved along a bit, then halted for about 5 hours. I rebooted my machine and tried the Xubuntu 10.10 installer WITHOUT the option to "download updates". The install completed in about 15 minutes. So, all of that is making me wonder if there is someting about 12.04 that does not like my hardware. I'm willing to try again, but only if I know I will not have to spend hours just to get to an error message and a hosed up system like I did last night. I also think I have a lot more RAM than is being reported in the output below. I had extra ram installed last year. I'm not good with the command line readouts, but there seems like there should be a lot more. I wasn't thrilled with Unity. I am willing to try Kubuntu 12.04. Will I run into the same problems? What is the highest version of a *ubuntu can I upgrade to? Thanks CPU Model: Intel(R) Pentium(R) 4 CPU 2.53GHz Frequency: 2533.223 MHz L2 Cache: 512 KB Bogomips: 5066.44 Numbering: family(15) model(2) stepping(7) Flags: fpu vme de pse tsc msr pae mce cx8 apic mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe up pebs bts cid RAM ~$ free -mt total used free shared buffers cached Mem: 1506 891 615 0 91 521 -/+ buffers/cache: 278 1227 Swap: 1609 0 1609 Total: 3116 891 2225 Video Card 01:00.0 VGA compatible controller: nVidia Corporation NV18 [GeForce4 MX 440 AGP 8x] (rev a2) (prog-if 00 [VGA controller]) Subsystem: ASUSTeK Computer Inc. V9180 Magic Flags: bus master, 66MHz, medium devsel, latency 32, IRQ 16 Memory at fd000000 (32-bit, non-prefetchable) [size=16M] Memory at f0000000 (32-bit, prefetchable) [size=64M] Expansion ROM at fe9e0000 [disabled] [size=128K] Kernel driver in use: nouveau Kernel modules: nouveau, nvidiafb Motherboard Intel 845PE ATX 533FSB DDR333 USB2

    Read the article

  • Battery doesn't charge on high cpu load

    - by bhappy
    When my cpu load rises the battery stop working I don't know why i.e If I start playing a game such as counter strike source the battery won't change unless i minimize the window which brings the load down Can someone please help me with this issue Note: sometimes when flash lags for a sec it shows discharging and charging again also due to high cpu load My laptop is a sony vaio F series 127FD model Thanks

    Read the article

  • Copying Tables from a Website

    - by amemus
    I have difficulty making an Excel-readable file from a table on a Website. The problems very specific to my question are: I have to use IE 7 to access the site. Excel is installed in another computer. The site does not let me view the HTML of the table. Normally, I would simply select the table I want and drag and drop it to Excel. Or, I would view the page source and copy the HTML data. Both do not work in this case. Is there any handy tool out there?

    Read the article

  • Why is my library/Documents shortcut showing all files from all subfolders in one flat list?

    - by korro
    I am using Windows Explorer on Windows 7. In My Documents there's a folder 02 KOOP. If I go there using the adress bar: D:\users\username\My Documents\02 KOOP I get a view of my subfolders and some files: However when I use the Documents shortcut under Libraries, and then go to the 02 KOOP folder, all files that are located in all subfolders are listed in one big list. When you view the file properties, the actual location is shown: This is the only subfolder of My Documents that has this behaviour. Did I accidentally change some setting, or is something wrong? Is this some setting I accidentally changed?

    Read the article

  • Sharepoint 2010 can't find domain users when granting permissions

    - by quani
    I'm trying to grant permissions to other people to view a SharePoint site but when granting permissions it uses "Check Names" and claims any user or group that is part of a domain does not exist. It does this if I try granting permissions to the team site or in central admin BUT if I try to add someone to Farm Administrators in Central admin then all of the sudden it can find all domain users. Why is it finding domain users in that one context but not others? It is supposed to be using NTLM authentication and has Windows configured as the authentication provider (And IIS is configured to use NTLM). What's even more strange is I enabled Anonymous Access for the team site which I thought would allow anyone to view it but others say they can't access it.

    Read the article

  • At what point does caching become necessary for a web application?

    - by Zaemz
    I'm considering the architecture for a web application. It's going to be a single page application that updates itself whenever the user selects different information on several forms that are available that are on the page. I was thinking that it shouldn't be good to rely on the user's browser to correctly interpret the information and update the view, so I'll send the user's choices to the server, and then get the data, send it back to the browser, and update the view. There's a table with 10,000 or so rows in a MySQL database that's going to be accessed pretty often, like once every 5-30 seconds for each user. I'm expecting 200-300 concurrent users at one time. I've read that a well designed relational database with simple queries are nothing for a RDBMS to handle, really, but I would still like to keep things quick for the client. Should this even be a concern for me at the moment? At what point would it be helpful to start using a separate caching service like Memcached or Redis, or would it even be necessary? I know that MySQL caches popular queries and the results, would this suffice?

    Read the article

  • Breathing for game/movie characters

    - by dtldarek
    Breathing (the movement of chest and face features): I'd like to ask if it is hard to model and whether it is computationaly expensive. I recently noticed the great effect it has in Madagascar 3 movie, but (please, correct me if I am wrong) don't remember seeing it in any games (except maybe steam cloud in cold/winter setting) and very few animated movies does that to noticable degree (e.g. when it is necessary by the plot or situation). I'd greatly appreciate answers from both movie graphics and game graphics perspective.

    Read the article

  • Unable to connect to sharedfax error

    - by Owl
    I'm running SBS2k on a machine with a 56k modem for all outgoing faxes. We have a terminal server running server 2003 enterprise and in printers the sharedfax shows "unable to connect status". When you right-click, properties, it gives you an error saying the spool service is not running and you cannot view properties. The spooler service is indeed running on the 2k3 server because I can view the other printer's properties..and on 2k I show the Microsoft Shared Fax service AND Print Spooler as started. Here's the tricky part, I can add the shared fax printer on a domain computer locally (Windows XP) and it will show ready status and I can send faxes, they show up on the 2k server's fax console manager as in progress. I've sent a couple from a few different workstations today successfully, but still unable to send from 2k3 terminal services.

    Read the article

  • Tool or website or process to display previews of website templates residing in archive files?

    - by Tony_Henrich
    I have hundreds of website templates in rar or zip files. To view any of them I have to extract the archive to a temporary folder and then view the template in there. It's a time consuming manual process to do this for each template Is there a tool which enables me to quickly preview the templates in the files? OR (if I extract each template into a separate folder off a master folder) A web app which can enable previewing of each template by automatically creating a link or a preview image (similar to template sites) of the home page for? OR any method to preview the templates in the fastest convenient way possible?

    Read the article

  • LEONARDI Application Suite 8.9 disponible, l'environnement open source de W4 pour les applications métier étend ses fonctionnalités

    LEONARDI Application Suite 8.9 disponible l'environnement open source de W4 pour les applications métier étend ses fonctionnalités LEONARDI Application Suite, la suite logicielle complète, fondée sur un framework open-source, pour créer et faire vivre des applications sur mesure suivant l'approche "Model Driven", évolue pour répondre aux besoins des utilisateurs. Application Suite 8.9, la dernière mise à jour de l'outil étend ses fonctionnalités au niveau conception, améliore l'ergonomie finale des applications, tout en renforçant leur qualité et en augmentant leur interopérabilité. Concernant l'aspect fonctionnel, le concepteur d'applications dispose d...

    Read the article

  • Specify Linq To SQL ConnectionString explicitly

    - by Michael Freidgeim
    When modifying Linq to  Sql data model in Visual Studio 2010,  it re-assigns ConnectionString that is available on developer’s machine. Because the name can be different on different machines, Designer often replace it with something like ConnectionString1, which causes errors during deployment.It requires developers to ensure that ConnectionString stays unchanged.  More reliable way is to use context constructor with explicit ConnectionString name instead of parameterless default constructor GOOD:   var ctx = new MyModelDataContext(Settings.Default.ConnectionString);Not good:          var ctx = new MyModelDataContext();

    Read the article

  • UCS Documentation: Home1 or Home2? You Decide.

    - by joesciallo
    How you go about finding information can be a very personal affair. We each have our own style of locating content, we each have our particular bent. But when it comes to getting started finding technical information, sometimes the simplest ways are best, especially if you are relatively new to a product or technology, and you just need to get going quickly with that Installation Guide or those Release Notes. With that in mind, I recently created an alternative home page for the Unified Communications Suite documentation. You can now have your Home2, in addition to the original, more wiki-fied Home1. I would recommend Home2 for those who are used to, and are more comfortable with, the spreadsheet-like view of manuals: here you'll find those familiar titles and the option to either view on the wiki or download the equivalent PDF file. Once you get familiar with what guides are available for the UCS component products, then move on to the Home1 layout, and start using some of the more advanced techniques for finding content in a wiki. Either way, you should be able to locate the "thing" you are looking for.

    Read the article

  • Is a 1:* write:read thread system safe?

    - by Di-0xide
    Theoretically, thread-safe code should fix race conditions. Race conditions, as I understand it, occur because two threads attempt to write to the same location at the same time. However, what about a threading model in which a single thread is designed to write to a location, and several slave/worker threads simply read from the location? Assuming the value/timing at which they read the data isn't relevant/doesn't hinder the worker thread's outcome, wouldn't this be considered 'thread safe', or am I missing something in my logic?

    Read the article

  • Installed Ubuntu on usb drive, now it's read-only and can't be formatted

    - by weiszam
    I installed Ubuntu (ubuntu-13.04-desktop-amd64) on a 8 gb pendrive. Because it worked pretty slow, I wanted to change it back to a normal pendrive. But this is what happened. Windows7 can't open the usb drive, can't view files or anything, can't access attributes. When I select to format, it says it can't format it because it is write-protected. Tried the same running from an Ubuntu, and trying from a booted GParted thing. When I view the partitions, they can't be deleted either. What should I do to get it formatted?

    Read the article

< Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >