Normally when you create AutoMapper profile, it is you responsibility to register them. However with the dependency injection container (DI) you have great possibility to automate this.
One of products I’m working on is using Telerik’s Kendo UI. It is nice, but sometimes not so intuitive and not very well documented, product suite.
Ok, don’t be cross with me, I know the comments aren’t clean code. But sometimes we need them. ReSharper templates is excellent possibility to write your comments quicker and personalized so you can ask the next time: “hey guy what did you meant with this?”.
Before I show you the cooperation between the Autofac and Microsoft Extensibility Framework (MEF) I want bring you in short my opinion why is there the place for both of them. If you are try to search around, you will find lot of posts which describes the differences between the MEF or Autofac, also, you will find lot of posts which advocates to use MEF as IoC Container, or, on other side, to use Autofac as replacement of the MEF.
Maybe you are surprised, maybe not, but fact is that you can increase the performance of reading and writing data with the explicit database connections.
There are many approaches how we can join the data in the memory with the data in database. But you should know the advantages and disadvantages for each one.
The performance measurement is one of the most important things in the software development. Performance should be addressed throughout the whole development process and not only when the risks become problems. Major motivation for running something in the cloud is the fault-tolerance and scalability. Notice: without performance, scalability doesn’t exists.
“Performance and scalability are inseparable from each other. It doesn’t make sense to talk about scalability if a software system doesn’t perform. However, a software system may perform but not scale.”
This article shows you an approach how to dynamically measure performance of your code at runtime. With the help of PostSharp I created the aspect which is applicable to the all members in the solution. This aspect is using the performance counter of type AverageTimer32 to the measure average execution time.
It is some time ago that is possible to use Virtual Machine(VM) in Windows Azure. A good example when you could use VM is build your own FTP Server hosted in Azure. The problem is that most products are using hard disk to store the data. The question is how you could access the data from your Web or Worker Role instances.
NuGet is just great tool for the managing third-party assemblies in your Visual Studio projects. If you are using the ReSharper and you are using a type that isn’t referenced in the current project but in another one, ReSharper will offer to add new reference just with the one click. In the case of NuGet, those assemblies are indeed referenced from the right location, but are not using the packages.config metadata file, which is essential for the NuGet to know which assemblies are referenced by what project.
But is seems there is already for a couple of time a solution for this. JetBrains (the company behind ReSharper) released plugins for ReSharper 7.1 and 8.0.
- 7.1 https://github.com/JetBrains/resharper-nuget/wiki/Release-Notes
- 8.0 http://resharper-plugins.jetbrains.com/packages/ReSharper.NuGet
If you want to know more details about the issue read here.
- Windows Azure: Just plug-in your work: Part 1
- Windows Azure: Just plug-in your work: Part 2
In my previous post I described the idea of using plug-ins in Windows Azure. In context of this sample we implemented a general plug-in framework. As we saw in the video, we need two services: one which is taking the work orders and stores them in the repository and the other one which is processing them.
WorkOrderGatewayService is representing a service which takes the work orders and stores them in the BLOB Storage. This service also answers the question whether the work order is done or not.
WorkOrderProcessorService is representing the service which is processing the work orders.